Warning: Permanently added '18.206.218.137' (ED25519) to the list of known hosts. You can reproduce this build on your computer by running: sudo dnf install copr-rpmbuild /usr/bin/copr-rpmbuild --verbose --drop-resultdir --task-url https://copr.fedorainfracloud.org/backend/get-build-task/8487194-fedora-41-x86_64 --chroot fedora-41-x86_64 Version: 1.2 PID: 17231 Logging PID: 17232 Task: {'allow_user_ssh': False, 'appstream': False, 'background': True, 'build_id': 8487194, 'buildroot_pkgs': [], 'chroot': 'fedora-41-x86_64', 'enable_net': False, 'fedora_review': False, 'git_hash': 'f67ee79ce262b7b6b38a2e8367bc43e8cd908af5', 'git_repo': 'https://copr-dist-git.fedorainfracloud.org/git/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch', 'isolation': 'default', 'memory_reqs': 2048, 'package_name': 'python-torch', 'package_version': '2.4.0-10', 'project_dirname': 'jinja2-3.1.5-f41:custom:isolated', 'project_name': 'jinja2-3.1.5-f41', 'project_owner': 'churchyard', 'repo_priority': None, 'repos': [{'baseurl': 'https://download.copr.fedorainfracloud.org/results/churchyard/jinja2-3.1.5-f41/fedora-41-x86_64/', 'id': 'copr_base', 'name': 'Copr repository', 'priority': None}, {'baseurl': 'http://kojipkgs.fedoraproject.org/repos/f41-build/latest/$basearch/', 'id': 'http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch', 'name': 'Additional repo http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch'}], 'sandbox': 'churchyard/jinja2-3.1.5-f41--churchyard', 'source_json': {}, 'source_type': None, 'ssh_public_keys': None, 'storage': 0, 'submitter': 'churchyard', 'tags': [], 'task_id': '8487194-fedora-41-x86_64', 'timeout': 18000, 'uses_devel_repo': False, 'with_opts': [], 'without_opts': []} Running: git clone https://copr-dist-git.fedorainfracloud.org/git/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch --depth 500 --no-single-branch --recursive cmd: ['git', 'clone', 'https://copr-dist-git.fedorainfracloud.org/git/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch', '/var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch', '--depth', '500', '--no-single-branch', '--recursive'] cwd: . rc: 0 stdout: stderr: Cloning into '/var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch'... Running: git checkout f67ee79ce262b7b6b38a2e8367bc43e8cd908af5 -- cmd: ['git', 'checkout', 'f67ee79ce262b7b6b38a2e8367bc43e8cd908af5', '--'] cwd: /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch rc: 0 stdout: stderr: Note: switching to 'f67ee79ce262b7b6b38a2e8367bc43e8cd908af5'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by switching back to a branch. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false HEAD is now at f67ee79 automatic import of python-torch Running: dist-git-client sources cmd: ['dist-git-client', 'sources'] cwd: /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch rc: 0 stdout: stderr: INFO: Reading stdout from command: git rev-parse --abbrev-ref HEAD INFO: Reading stdout from command: git rev-parse HEAD INFO: Reading sources specification file: sources INFO: Downloading cpp-httplib-3b6597b.tar.gz INFO: Reading stdout from command: curl --help all INFO: Calling: curl -H Pragma: -o cpp-httplib-3b6597b.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/cpp-httplib-3b6597b.tar.gz/md5/8fe49d994d349a1f29f327eba8beeefc/cpp-httplib-3b6597b.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 1129k 100 1129k 0 0 93.2M 0 --:--:-- --:--:-- --:--:-- 100M INFO: Reading stdout from command: md5sum cpp-httplib-3b6597b.tar.gz INFO: Downloading kineto-be13176.tar.gz INFO: Calling: curl -H Pragma: -o kineto-be13176.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/kineto-be13176.tar.gz/md5/9f82e0992eeaa11c940f3fee6add932b/kineto-be13176.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 8050k 100 8050k 0 0 345M 0 --:--:-- --:--:-- --:--:-- 357M INFO: Reading stdout from command: md5sum kineto-be13176.tar.gz INFO: Downloading libnop-910b558.tar.gz INFO: Calling: curl -H Pragma: -o libnop-910b558.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/libnop-910b558.tar.gz/md5/6a346d3173f36e0807f8885e096fc3ce/libnop-910b558.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 121k 100 121k 0 0 11.7M 0 --:--:-- --:--:-- --:--:-- 13.1M /usr/bin/tail: /var/lib/copr-rpmbuild/main.log: file truncated INFO: Reading stdout from command: md5sum libnop-910b558.tar.gz INFO: Downloading pytorch-v2.4.0.tar.gz INFO: Calling: curl -H Pragma: -o pytorch-v2.4.0.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/pytorch-v2.4.0.tar.gz/md5/8604669939cd671b648f2280139dc242/pytorch-v2.4.0.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 283M 100 283M 0 0 559M 0 --:--:-- --:--:-- --:--:-- 558M INFO: Reading stdout from command: md5sum pytorch-v2.4.0.tar.gz INFO: Downloading tensorpipe-52791a2.tar.gz INFO: Calling: curl -H Pragma: -o tensorpipe-52791a2.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/tensorpipe-52791a2.tar.gz/md5/8d2ad81f75a80fdafbbd51600a2617da/tensorpipe-52791a2.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 253k 100 253k 0 0 23.5M 0 --:--:-- --:--:-- --:--:-- 24.7M INFO: Reading stdout from command: md5sum tensorpipe-52791a2.tar.gz INFO: Downloading v1.14.2.tar.gz INFO: Calling: curl -H Pragma: -o v1.14.2.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/v1.14.2.tar.gz/md5/b56fb37cc0c9ca5909e46487906fecee/v1.14.2.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 799k 100 799k 0 0 68.2M 0 --:--:-- --:--:-- --:--:-- 70.9M INFO: Reading stdout from command: md5sum v1.14.2.tar.gz INFO: Downloading v1.41.0.tar.gz INFO: Calling: curl -H Pragma: -o v1.41.0.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/v1.41.0.tar.gz/md5/1423d5639279fa7ed58a3420c4eb811f/v1.41.0.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 1251k 100 1251k 0 0 99.6M 0 --:--:-- --:--:-- --:--:-- 101M INFO: Reading stdout from command: md5sum v1.41.0.tar.gz INFO: Downloading v2.11.1.tar.gz INFO: Calling: curl -H Pragma: -o v2.11.1.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/v2.11.1.tar.gz/md5/49e92f92244021912a56935918c927d0/v2.11.1.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 738k 100 738k 0 0 60.4M 0 --:--:-- --:--:-- --:--:-- 65.5M INFO: Reading stdout from command: md5sum v2.11.1.tar.gz INFO: Downloading v23.3.3.tar.gz INFO: Calling: curl -H Pragma: -o v23.3.3.tar.gz --location --connect-timeout 60 --retry 3 --retry-delay 10 --remote-time --show-error --fail --retry-all-errors https://copr-dist-git.fedorainfracloud.org/repo/pkgs/churchyard/jinja2-3.1.5-f41:custom:isolated/python-torch/v23.3.3.tar.gz/md5/a379be2558d12c02acdb062ad7c8969c/v23.3.3.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 2145k 100 2145k 0 0 156M 0 --:--:-- --:--:-- --:--:-- 161M INFO: Reading stdout from command: md5sum v23.3.3.tar.gz Running (timeout=18000): unbuffer mock --spec /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch/python-torch.spec --sources /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch --resultdir /var/lib/copr-rpmbuild/results --uniqueext 1736346719.362177 -r /var/lib/copr-rpmbuild/results/configs/child.cfg INFO: mock.py version 6.0 starting (python version = 3.13.0, NVR = mock-6.0-1.fc41), args: /usr/libexec/mock/mock --spec /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch/python-torch.spec --sources /var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch --resultdir /var/lib/copr-rpmbuild/results --uniqueext 1736346719.362177 -r /var/lib/copr-rpmbuild/results/configs/child.cfg Start(bootstrap): init plugins INFO: tmpfs initialized INFO: selinux enabled INFO: chroot_scan: initialized INFO: compress_logs: initialized Finish(bootstrap): init plugins Start: init plugins INFO: tmpfs initialized INFO: selinux enabled INFO: chroot_scan: initialized INFO: compress_logs: initialized Finish: init plugins INFO: Signal handler active Start: run INFO: Start(/var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch/python-torch.spec) Config(fedora-41-x86_64) Start: clean chroot Finish: clean chroot Mock Version: 6.0 INFO: Mock Version: 6.0 Start(bootstrap): chroot init INFO: mounting tmpfs at /var/lib/mock/fedora-41-x86_64-bootstrap-1736346719.362177/root. INFO: calling preinit hooks INFO: enabled root cache INFO: enabled package manager cache Start(bootstrap): cleaning package manager metadata Finish(bootstrap): cleaning package manager metadata INFO: Guessed host environment type: unknown INFO: Using container image: registry.fedoraproject.org/fedora:41 INFO: Pulling image: registry.fedoraproject.org/fedora:41 INFO: Tagging container image as mock-bootstrap-1cbb52b7-a088-4e66-a70f-16aebea598c9 INFO: Checking that 4b6e09cb21b0b21e79a54d011edfb4c7969b61b24323a64261cbe34bd294d49e image matches host's architecture INFO: Copy content of container 4b6e09cb21b0b21e79a54d011edfb4c7969b61b24323a64261cbe34bd294d49e to /var/lib/mock/fedora-41-x86_64-bootstrap-1736346719.362177/root INFO: mounting 4b6e09cb21b0b21e79a54d011edfb4c7969b61b24323a64261cbe34bd294d49e with podman image mount INFO: image 4b6e09cb21b0b21e79a54d011edfb4c7969b61b24323a64261cbe34bd294d49e as /var/lib/containers/storage/overlay/8f016905ed6520775034a1d28a2ef6287295c05be42b7f9fb8c7a60231d9de29/merged INFO: umounting image 4b6e09cb21b0b21e79a54d011edfb4c7969b61b24323a64261cbe34bd294d49e (/var/lib/containers/storage/overlay/8f016905ed6520775034a1d28a2ef6287295c05be42b7f9fb8c7a60231d9de29/merged) with podman image umount INFO: Removing image mock-bootstrap-1cbb52b7-a088-4e66-a70f-16aebea598c9 INFO: Package manager dnf5 detected and used (fallback) INFO: Not updating bootstrap chroot, bootstrap_image_ready=True Start(bootstrap): creating root cache Finish(bootstrap): creating root cache Finish(bootstrap): chroot init Start: chroot init INFO: mounting tmpfs at /var/lib/mock/fedora-41-x86_64-1736346719.362177/root. INFO: calling preinit hooks INFO: enabled root cache INFO: enabled package manager cache Start: cleaning package manager metadata Finish: cleaning package manager metadata INFO: enabled HW Info plugin INFO: Package manager dnf5 detected and used (direct choice) INFO: Buildroot is handled by package management downloaded with a bootstrap image: rpm-4.20.0-1.fc41.x86_64 rpm-sequoia-1.7.0-3.fc41.x86_64 dnf5-5.2.8.1-2.fc41.x86_64 dnf5-plugins-5.2.8.1-2.fc41.x86_64 Start: installing minimal buildroot with dnf5 Updating and loading repositories: updates 100% | 180.4 KiB/s | 22.4 KiB | 00m00s fedora 100% | 495.5 KiB/s | 29.7 KiB | 00m00s Copr repository 100% | 72.1 KiB/s | 1.5 KiB | 00m00s Additional repo http_kojipkgs_fedorapr 100% | 201.6 KiB/s | 3.8 KiB | 00m00s Repositories loaded. Package Arch Version Repository Size Installing group/module packages: bash x86_64 5.2.32-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.2 MiB bzip2 x86_64 1.0.8-19.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 95.7 KiB coreutils x86_64 9.5-11.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.7 MiB cpio x86_64 2.15-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.1 MiB diffutils x86_64 3.10-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.6 MiB fedora-release-common noarch 41-29 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 19.7 KiB findutils x86_64 1:4.10.0-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 MiB gawk x86_64 5.3.0-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB glibc-minimal-langpack x86_64 2.40-17.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B grep x86_64 3.11-9.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 MiB gzip x86_64 1.13-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 389.0 KiB info x86_64 7.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 361.8 KiB patch x86_64 2.7.6-25.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 266.7 KiB redhat-rpm-config noarch 293-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 183.5 KiB rpm-build x86_64 4.20.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 194.3 KiB sed x86_64 4.9-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 861.5 KiB shadow-utils x86_64 2:4.15.1-12.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.1 MiB tar x86_64 2:1.35-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.9 MiB unzip x86_64 6.0-64.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 386.8 KiB util-linux x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.7 MiB which x86_64 2.21-42.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 80.2 KiB xz x86_64 1:5.6.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.2 MiB Installing dependencies: add-determinism x86_64 0.3.6-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.4 MiB alternatives x86_64 1.31-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 64.8 KiB ansible-srpm-macros noarch 1-16.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 35.7 KiB audit-libs x86_64 4.0.2-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 331.3 KiB authselect x86_64 1.5.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 157.6 KiB authselect-libs x86_64 1.5.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 822.2 KiB basesystem noarch 11-21.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B binutils x86_64 2.43.1-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 27.4 MiB build-reproducibility-srpm-macros noarch 0.3.6-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 735.0 B bzip2-libs x86_64 1.0.8-19.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 80.7 KiB ca-certificates noarch 2024.2.69_v8.0.401-1.0.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.4 MiB coreutils-common x86_64 9.5-11.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 11.2 MiB cracklib x86_64 2.9.11-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 238.9 KiB crypto-policies noarch 20241029-1.git8baf557.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 136.9 KiB curl x86_64 8.9.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 796.2 KiB cyrus-sasl-lib x86_64 2.1.28-27.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.3 MiB debugedit x86_64 5.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 200.6 KiB dwz x86_64 0.15-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 298.9 KiB ed x86_64 1.20.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 146.9 KiB efi-srpm-macros noarch 5-13.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 40.2 KiB elfutils x86_64 0.192-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.6 MiB elfutils-debuginfod-client x86_64 0.192-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 81.4 KiB elfutils-default-yama-scope noarch 0.192-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 KiB elfutils-libelf x86_64 0.192-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.2 MiB elfutils-libs x86_64 0.192-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 654.9 KiB fedora-gpg-keys noarch 41-1 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 126.4 KiB fedora-release noarch 41-29 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B fedora-release-identity-basic noarch 41-29 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 682.0 B fedora-repos noarch 41-1 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.9 KiB file x86_64 5.45-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 103.5 KiB file-libs x86_64 5.45-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 9.9 MiB filesystem x86_64 3.18-23.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 106.0 B fonts-srpm-macros noarch 1:2.0.5-17.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 55.8 KiB forge-srpm-macros noarch 0.4.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 38.9 KiB fpc-srpm-macros noarch 1.3-13.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 144.0 B gdb-minimal x86_64 15.2-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 13.0 MiB gdbm x86_64 1:1.23-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 460.9 KiB gdbm-libs x86_64 1:1.23-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 121.9 KiB ghc-srpm-macros noarch 1.9.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 747.0 B glibc x86_64 2.40-17.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 6.6 MiB glibc-common x86_64 2.40-17.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 MiB glibc-gconv-extra x86_64 2.40-17.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 7.8 MiB gmp x86_64 1:6.3.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 811.4 KiB gnat-srpm-macros noarch 6-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 KiB go-srpm-macros noarch 3.6.0-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 60.8 KiB jansson x86_64 2.13.1-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 88.3 KiB json-c x86_64 0.17-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 82.4 KiB kernel-srpm-macros noarch 1.0-24.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.9 KiB keyutils-libs x86_64 1.6.3-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 54.4 KiB krb5-libs x86_64 1.21.3-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.3 MiB libacl x86_64 2.3.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 40.0 KiB libarchive x86_64 3.7.4-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 926.6 KiB libattr x86_64 2.5.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 28.5 KiB libblkid x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 258.5 KiB libbrotli x86_64 1.1.0-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 837.6 KiB libcap x86_64 2.70-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 220.2 KiB libcap-ng x86_64 0.8.5-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 69.2 KiB libcom_err x86_64 1.47.1-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 67.2 KiB libcurl x86_64 8.9.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 818.1 KiB libeconf x86_64 0.6.2-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 58.0 KiB libevent x86_64 2.1.12-14.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 895.7 KiB libfdisk x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 362.9 KiB libffi x86_64 3.4.6-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 86.4 KiB libgcc x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 274.6 KiB libgomp x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 523.5 KiB libidn2 x86_64 2.3.7-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 329.1 KiB libmount x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 351.8 KiB libnghttp2 x86_64 1.62.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 166.1 KiB libnsl2 x86_64 2.0.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 57.9 KiB libpkgconf x86_64 2.3.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 78.2 KiB libpsl x86_64 0.21.5-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 80.5 KiB libpwquality x86_64 1.4.5-11.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 417.8 KiB libselinux x86_64 3.7-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 181.0 KiB libsemanage x86_64 3.7-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 293.5 KiB libsepol x86_64 3.7-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 817.8 KiB libsmartcols x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 180.4 KiB libssh x86_64 0.10.6-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 513.3 KiB libssh-config noarch 0.10.6-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 277.0 B libstdc++ x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.8 MiB libtasn1 x86_64 4.19.0-9.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 175.7 KiB libtirpc x86_64 1.3.6-1.rc3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 197.6 KiB libtool-ltdl x86_64 2.4.7-12.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 66.2 KiB libunistring x86_64 1.1-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB libutempter x86_64 1.2.1-15.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 57.7 KiB libuuid x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 37.5 KiB libverto x86_64 0.3.2-9.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 29.5 KiB libxcrypt x86_64 4.4.37-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 269.6 KiB libxml2 x86_64 2.12.9-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB libzstd x86_64 1.5.6-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 795.9 KiB lua-libs x86_64 5.4.6-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 285.0 KiB lua-srpm-macros noarch 1-14.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.3 KiB lz4-libs x86_64 1.10.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 145.5 KiB mpfr x86_64 4.2.1-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 832.1 KiB ncurses-base noarch 6.5-2.20240629.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 326.3 KiB ncurses-libs x86_64 6.5-2.20240629.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 975.2 KiB ocaml-srpm-macros noarch 10-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.9 KiB openblas-srpm-macros noarch 2-18.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 112.0 B openldap x86_64 2.6.8-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 635.4 KiB openssl-libs x86_64 1:3.2.2-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 7.7 MiB p11-kit x86_64 0.25.5-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.2 MiB p11-kit-trust x86_64 0.25.5-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 391.4 KiB package-notes-srpm-macros noarch 0.5-12.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.6 KiB pam x86_64 1.6.1-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 MiB pam-libs x86_64 1.6.1-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 139.0 KiB pcre2 x86_64 10.44-1.fc41.1 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 653.5 KiB pcre2-syntax noarch 10.44-1.fc41.1 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 251.6 KiB perl-srpm-macros noarch 1-56.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 861.0 B pkgconf x86_64 2.3.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 88.6 KiB pkgconf-m4 noarch 2.3.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 14.4 KiB pkgconf-pkg-config x86_64 2.3.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 989.0 B popt x86_64 1.19-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 136.9 KiB publicsuffix-list-dafsa noarch 20240107-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 67.5 KiB pyproject-srpm-macros noarch 1.16.3-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.9 KiB python-srpm-macros noarch 3.13-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 51.0 KiB qt5-srpm-macros noarch 5.15.15-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 500.0 B qt6-srpm-macros noarch 6.8.1-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 456.0 B readline x86_64 8.2-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 493.2 KiB rpm x86_64 4.20.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.1 MiB rpm-build-libs x86_64 4.20.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 206.7 KiB rpm-libs x86_64 4.20.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 725.9 KiB rpm-sequoia x86_64 1.7.0-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.3 MiB rust-srpm-macros noarch 26.3-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.8 KiB setup noarch 2.15.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 720.7 KiB sqlite-libs x86_64 3.46.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.4 MiB systemd-libs x86_64 256.10-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.0 MiB util-linux-core x86_64 2.40.2-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.5 MiB xxhash-libs x86_64 0.8.3-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 88.5 KiB xz-libs x86_64 1:5.6.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 214.4 KiB zig-srpm-macros noarch 1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.1 KiB zip x86_64 3.0-41.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 703.2 KiB zlib-ng-compat x86_64 2.1.7-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 134.0 KiB zstd x86_64 1.5.6-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB Installing groups: Buildsystem building group Transaction Summary: Installing: 154 packages Total size of inbound packages is 52 MiB. Need to download 0 B. After this operation, 180 MiB extra will be used (install 180 MiB, remove 0 B). [1/1] tar-2:1.35-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [1/1] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/2] bzip2-0:1.0.8-19.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [2/2] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/3] redhat-rpm-config-0:293-1.fc41.no 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [3/3] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/4] rpm-build-0:4.20.0-1.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [4/4] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/5] unzip-0:6.0-64.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [5/5] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/6] cpio-0:2.15-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [6/6] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/7] which-0:2.21-42.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [7/7] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/8] bash-0:5.2.32-1.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [8/8] Total 100% | 0.0 B/s | 0.0 B | 00m00s [1/9] coreutils-0:9.5-11.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [9/9] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/10] grep-0:3.11-9.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [10/10] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/11] patch-0:2.7.6-25.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [11/11] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/12] sed-0:4.9-3.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [12/12] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/13] shadow-utils-2:4.15.1-12.fc41.x 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [13/13] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/14] util-linux-0:2.40.2-4.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [14/14] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/15] diffutils-0:3.10-8.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [15/15] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/16] fedora-release-common-0:41-29.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [16/16] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/17] findutils-1:4.10.0-4.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [17/17] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/18] gawk-0:5.3.0-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [18/18] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/19] glibc-minimal-langpack-0:2.40-1 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [19/19] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/20] gzip-0:1.13-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [20/20] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/21] info-0:7.1-3.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [21/21] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/22] xz-1:5.6.2-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [22/22] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/23] glibc-0:2.40-17.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [23/23] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/24] xz-libs-1:5.6.2-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [24/24] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/25] pam-0:1.6.1-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [25/25] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/26] pam-libs-0:1.6.1-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [26/26] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/27] systemd-libs-0:256.10-1.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [27/27] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/28] audit-libs-0:4.0.2-1.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [28/28] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/29] authselect-libs-0:1.5.0-8.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [29/29] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/30] libblkid-0:2.40.2-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [30/30] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/31] libcap-ng-0:0.8.5-3.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [31/31] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/32] libfdisk-0:2.40.2-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [32/32] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/33] libmount-0:2.40.2-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [33/33] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/34] libselinux-0:3.7-5.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [34/34] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/35] libsmartcols-0:2.40.2-4.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [35/35] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/36] libutempter-0:1.2.1-15.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [36/36] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/37] libuuid-0:2.40.2-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [37/37] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/38] ncurses-libs-0:6.5-2.20240629.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [38/38] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/39] readline-0:8.2-10.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [39/39] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/40] util-linux-core-0:2.40.2-4.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [40/40] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/41] zlib-ng-compat-0:2.1.7-3.fc41.x 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [41/41] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/42] bzip2-libs-0:1.0.8-19.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [42/42] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/43] libacl-0:2.3.2-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [43/43] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/44] libcap-0:2.70-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [44/44] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/45] libgcc-0:14.2.1-3.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [45/45] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/46] libxcrypt-0:4.4.37-4.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [46/46] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/47] setup-0:2.15.0-8.fc41.noarch 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [47/47] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/48] libattr-0:2.5.2-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [48/48] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/49] libeconf-0:0.6.2-3.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [49/49] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/50] libsemanage-0:3.7-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [50/50] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/51] binutils-0:2.43.1-5.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [51/51] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/52] debugedit-0:5.1-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [52/52] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/53] elfutils-0:0.192-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [53/53] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/54] elfutils-libelf-0:0.192-7.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [54/54] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/55] libarchive-0:3.7.4-4.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [55/55] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/56] file-0:5.45-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [56/56] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/57] pkgconf-pkg-config-0:2.3.0-1.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [57/57] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/58] popt-0:1.19-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [58/58] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/59] rpm-0:4.20.0-1.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [59/59] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/60] rpm-build-libs-0:4.20.0-1.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [60/60] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/61] rpm-libs-0:4.20.0-1.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [61/61] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/62] zstd-0:1.5.6-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [62/62] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/63] curl-0:8.9.1-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [63/63] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/64] glibc-gconv-extra-0:2.40-17.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [64/64] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/65] build-reproducibility-srpm-macr 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [65/65] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/66] efi-srpm-macros-0:5-13.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [66/66] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/67] forge-srpm-macros-0:0.4.0-1.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [67/67] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/68] go-srpm-macros-0:3.6.0-5.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [68/68] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/69] pyproject-srpm-macros-0:1.16.3- 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [69/69] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/70] qt6-srpm-macros-0:6.8.1-4.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [70/70] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/71] ansible-srpm-macros-0:1-16.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [71/71] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/72] dwz-0:0.15-8.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [72/72] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/73] fonts-srpm-macros-1:2.0.5-17.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [73/73] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/74] fpc-srpm-macros-0:1.3-13.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [74/74] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/75] ghc-srpm-macros-0:1.9.1-2.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [75/75] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/76] gnat-srpm-macros-0:6-6.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [76/76] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/77] kernel-srpm-macros-0:1.0-24.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [77/77] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/78] lua-srpm-macros-0:1-14.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [78/78] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/79] ocaml-srpm-macros-0:10-3.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [79/79] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/80] openblas-srpm-macros-0:2-18.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [80/80] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/81] package-notes-srpm-macros-0:0.5 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [81/81] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/82] perl-srpm-macros-0:1-56.fc41.no 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [82/82] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/83] python-srpm-macros-0:3.13-3.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [83/83] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/84] qt5-srpm-macros-0:5.15.15-1.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [84/84] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/85] rust-srpm-macros-0:26.3-3.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [85/85] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/86] zig-srpm-macros-0:1-3.fc41.noar 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [86/86] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/87] zip-0:3.0-41.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [87/87] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/88] pkgconf-0:2.3.0-1.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [88/88] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/89] pkgconf-m4-0:2.3.0-1.fc41.noarc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [89/89] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/90] libpkgconf-0:2.3.0-1.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [90/90] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/91] ed-0:1.20.2-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [91/91] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/92] libtirpc-0:1.3.6-1.rc3.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [92/92] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/93] openssl-libs-1:3.2.2-10.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [93/93] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/94] authselect-0:1.5.0-8.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [94/94] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/95] gdbm-1:1.23-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [95/95] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/96] gdbm-libs-1:1.23-7.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [96/96] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/97] libnsl2-0:2.0.1-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [97/97] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/98] libpwquality-0:1.4.5-11.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [98/98] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/99] crypto-policies-0:20241029-1.gi 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [99/99] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/100] ca-certificates-0:2024.2.69_v 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [100/100] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/101] ncurses-base-0:6.5-2.20240629 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [101/101] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/102] krb5-libs-0:1.21.3-3.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [102/102] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/103] libcom_err-0:1.47.1-6.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [103/103] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/104] libsepol-0:3.7-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [104/104] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/105] pcre2-0:10.44-1.fc41.1.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [105/105] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/106] cracklib-0:2.9.11-6.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [106/106] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/107] libxml2-0:2.12.9-1.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [107/107] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/108] libzstd-0:1.5.6-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [108/108] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/109] lz4-libs-0:1.10.0-1.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [109/109] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/110] keyutils-libs-0:1.6.3-4.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [110/110] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/111] libverto-0:0.3.2-9.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [111/111] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/112] glibc-common-0:2.40-17.fc41.x 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [112/112] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/113] basesystem-0:11-21.fc41.noarc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [113/113] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/114] filesystem-0:3.18-23.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [114/114] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/115] gmp-1:6.3.0-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [115/115] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/116] mpfr-0:4.2.1-5.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [116/116] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/117] file-libs-0:5.45-7.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [117/117] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/118] fedora-repos-0:41-1.noarch 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [118/118] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/119] elfutils-debuginfod-client-0: 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [119/119] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/120] elfutils-libs-0:0.192-7.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [120/120] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/121] libstdc++-0:14.2.1-3.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [121/121] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/122] coreutils-common-0:9.5-11.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [122/122] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/123] libffi-0:3.4.6-3.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [123/123] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/124] p11-kit-0:0.25.5-3.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [124/124] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/125] p11-kit-trust-0:0.25.5-3.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [125/125] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/126] add-determinism-0:0.3.6-3.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [126/126] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/127] alternatives-0:1.31-1.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [127/127] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/128] jansson-0:2.13.1-10.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [128/128] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/129] rpm-sequoia-0:1.7.0-3.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [129/129] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/130] lua-libs-0:5.4.6-6.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [130/130] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/131] sqlite-libs-0:3.46.1-1.fc41.x 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [131/131] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/132] libgomp-0:14.2.1-3.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [132/132] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/133] pcre2-syntax-0:10.44-1.fc41.1 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [133/133] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/134] libtasn1-0:4.19.0-9.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [134/134] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/135] fedora-gpg-keys-0:41-1.noarch 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [135/135] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/136] elfutils-default-yama-scope-0 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [136/136] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/137] json-c-0:0.17-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [137/137] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/138] fedora-release-0:41-29.noarch 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [138/138] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/139] gdb-minimal-0:15.2-3.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [139/139] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/140] xxhash-libs-0:0.8.3-1.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [140/140] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/141] fedora-release-identity-basic 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [141/141] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/142] libcurl-0:8.9.1-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [142/142] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/143] openldap-0:2.6.8-6.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [143/143] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/144] libbrotli-0:1.1.0-5.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [144/144] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/145] libidn2-0:2.3.7-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [145/145] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/146] libnghttp2-0:1.62.1-2.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [146/146] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/147] libpsl-0:0.21.5-4.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [147/147] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/148] libssh-0:0.10.6-8.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [148/148] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/149] cyrus-sasl-lib-0:2.1.28-27.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [149/149] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/150] libevent-0:2.1.12-14.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [150/150] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/151] libtool-ltdl-0:2.4.7-12.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [151/151] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/152] libssh-config-0:0.10.6-8.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [152/152] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/153] libunistring-0:1.1-8.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [153/153] Total 100% | 0.0 B/s | 0.0 B | 00m00s [ 1/154] publicsuffix-list-dafsa-0:202 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded -------------------------------------------------------------------------------- [154/154] Total 100% | 0.0 B/s | 0.0 B | 00m00s Running transaction [ 1/156] Verify package files 100% | 922.0 B/s | 154.0 B | 00m00s [ 2/156] Prepare transaction 100% | 4.0 KiB/s | 154.0 B | 00m00s [ 3/156] Installing libgcc-0:14.2.1-3. 100% | 269.8 MiB/s | 276.3 KiB | 00m00s [ 4/156] Installing publicsuffix-list- 100% | 0.0 B/s | 68.3 KiB | 00m00s [ 5/156] Installing libssh-config-0:0. 100% | 0.0 B/s | 816.0 B | 00m00s [ 6/156] Installing fedora-release-ide 100% | 0.0 B/s | 940.0 B | 00m00s [ 7/156] Installing fedora-gpg-keys-0: 100% | 56.1 MiB/s | 172.2 KiB | 00m00s [ 8/156] Installing fedora-repos-0:41- 100% | 0.0 B/s | 5.7 KiB | 00m00s [ 9/156] Installing fedora-release-com 100% | 23.4 MiB/s | 24.0 KiB | 00m00s [ 10/156] Installing fedora-release-0:4 100% | 0.0 B/s | 124.0 B | 00m00s [ 11/156] Installing setup-0:2.15.0-8.f 100% | 64.5 MiB/s | 726.5 KiB | 00m00s >>> [RPM] /etc/hosts created as /etc/hosts.rpmnew [ 12/156] Installing filesystem-0:3.18- 100% | 3.6 MiB/s | 212.5 KiB | 00m00s [ 13/156] Installing basesystem-0:11-21 100% | 0.0 B/s | 124.0 B | 00m00s [ 14/156] Installing pcre2-syntax-0:10. 100% | 248.1 MiB/s | 254.1 KiB | 00m00s [ 15/156] Installing coreutils-common-0 100% | 373.0 MiB/s | 11.2 MiB | 00m00s [ 16/156] Installing ncurses-base-0:6.5 100% | 85.9 MiB/s | 351.7 KiB | 00m00s [ 17/156] Installing glibc-minimal-lang 100% | 0.0 B/s | 124.0 B | 00m00s [ 18/156] Installing ncurses-libs-0:6.5 100% | 239.7 MiB/s | 981.8 KiB | 00m00s [ 19/156] Installing glibc-0:2.40-17.fc 100% | 330.7 MiB/s | 6.6 MiB | 00m00s [ 20/156] Installing bash-0:5.2.32-1.fc 100% | 408.5 MiB/s | 8.2 MiB | 00m00s [ 21/156] Installing glibc-common-0:2.4 100% | 207.1 MiB/s | 1.0 MiB | 00m00s [ 22/156] Installing glibc-gconv-extra- 100% | 271.1 MiB/s | 7.9 MiB | 00m00s [ 23/156] Installing zlib-ng-compat-0:2 100% | 131.6 MiB/s | 134.8 KiB | 00m00s [ 24/156] Installing xz-libs-1:5.6.2-2. 100% | 210.4 MiB/s | 215.5 KiB | 00m00s [ 25/156] Installing bzip2-libs-0:1.0.8 100% | 0.0 B/s | 81.8 KiB | 00m00s [ 26/156] Installing libuuid-0:2.40.2-4 100% | 0.0 B/s | 38.6 KiB | 00m00s [ 27/156] Installing readline-0:8.2-10. 100% | 241.8 MiB/s | 495.3 KiB | 00m00s [ 28/156] Installing popt-0:1.19-7.fc41 100% | 70.1 MiB/s | 143.5 KiB | 00m00s [ 29/156] Installing libblkid-0:2.40.2- 100% | 253.4 MiB/s | 259.5 KiB | 00m00s [ 30/156] Installing libxcrypt-0:4.4.37 100% | 265.9 MiB/s | 272.3 KiB | 00m00s [ 31/156] Installing libattr-0:2.5.2-4. 100% | 0.0 B/s | 29.5 KiB | 00m00s [ 32/156] Installing libacl-0:2.3.2-2.f 100% | 0.0 B/s | 40.7 KiB | 00m00s [ 33/156] Installing libzstd-0:1.5.6-2. 100% | 389.3 MiB/s | 797.2 KiB | 00m00s [ 34/156] Installing elfutils-libelf-0: 100% | 389.0 MiB/s | 1.2 MiB | 00m00s [ 35/156] Installing gmp-1:6.3.0-2.fc41 100% | 397.3 MiB/s | 813.7 KiB | 00m00s [ 36/156] Installing libstdc++-0:14.2.1 100% | 395.2 MiB/s | 2.8 MiB | 00m00s [ 37/156] Installing libeconf-0:0.6.2-3 100% | 0.0 B/s | 59.7 KiB | 00m00s [ 38/156] Installing gdbm-libs-1:1.23-7 100% | 0.0 B/s | 123.6 KiB | 00m00s [ 39/156] Installing mpfr-0:4.2.1-5.fc4 100% | 271.4 MiB/s | 833.7 KiB | 00m00s [ 40/156] Installing gawk-0:5.3.0-4.fc4 100% | 288.7 MiB/s | 1.7 MiB | 00m00s [ 41/156] Installing dwz-0:0.15-8.fc41. 100% | 293.3 MiB/s | 300.3 KiB | 00m00s [ 42/156] Installing unzip-0:6.0-64.fc4 100% | 190.6 MiB/s | 390.3 KiB | 00m00s [ 43/156] Installing file-libs-0:5.45-7 100% | 709.6 MiB/s | 9.9 MiB | 00m00s [ 44/156] Installing file-0:5.45-7.fc41 100% | 17.1 MiB/s | 105.0 KiB | 00m00s >>> Running pre-install scriptlet: crypto-policies-0:20241029-1.git8baf557.fc41. >>> Finished pre-install scriptlet: crypto-policies-0:20241029-1.git8baf557.fc41 >>> Scriptlet output: >>> /var/tmp/rpm-tmp.jKEl6N: line 2: rm: command not found >>> [ 45/156] Installing crypto-policies-0: 100% | 31.9 MiB/s | 163.3 KiB | 00m00s [ 46/156] Installing libcap-ng-0:0.8.5- 100% | 69.4 MiB/s | 71.0 KiB | 00m00s [ 47/156] Installing audit-libs-0:4.0.2 100% | 325.6 MiB/s | 333.4 KiB | 00m00s [ 48/156] Installing pam-libs-0:1.6.1-7 100% | 137.9 MiB/s | 141.3 KiB | 00m00s [ 49/156] Installing libcap-0:2.70-4.fc 100% | 110.0 MiB/s | 225.2 KiB | 00m00s [ 50/156] Installing systemd-libs-0:256 100% | 331.6 MiB/s | 2.0 MiB | 00m00s [ 51/156] Installing libsmartcols-0:2.4 100% | 177.1 MiB/s | 181.4 KiB | 00m00s [ 52/156] Installing libcom_err-0:1.47. 100% | 0.0 B/s | 68.3 KiB | 00m00s [ 53/156] Installing libsepol-0:3.7-2.f 100% | 399.8 MiB/s | 818.8 KiB | 00m00s [ 54/156] Installing pcre2-0:10.44-1.fc 100% | 319.8 MiB/s | 654.9 KiB | 00m00s [ 55/156] Installing libselinux-0:3.7-5 100% | 178.0 MiB/s | 182.3 KiB | 00m00s [ 56/156] Installing sed-0:4.9-3.fc41.x 100% | 212.3 MiB/s | 869.7 KiB | 00m00s [ 57/156] Installing grep-0:3.11-9.fc41 100% | 250.8 MiB/s | 1.0 MiB | 00m00s [ 58/156] Installing findutils-1:4.10.0 100% | 309.7 MiB/s | 1.9 MiB | 00m00s [ 59/156] Installing xz-1:5.6.2-2.fc41. 100% | 241.0 MiB/s | 1.2 MiB | 00m00s [ 60/156] Installing libmount-0:2.40.2- 100% | 344.7 MiB/s | 352.9 KiB | 00m00s [ 61/156] Installing lz4-libs-0:1.10.0- 100% | 143.1 MiB/s | 146.6 KiB | 00m00s [ 62/156] Installing libffi-0:3.4.6-3.f 100% | 85.7 MiB/s | 87.8 KiB | 00m00s [ 63/156] Installing alternatives-0:1.3 100% | 0.0 B/s | 66.4 KiB | 00m00s [ 64/156] Installing lua-libs-0:5.4.6-6 100% | 279.5 MiB/s | 286.2 KiB | 00m00s [ 65/156] Installing libtasn1-0:4.19.0- 100% | 173.3 MiB/s | 177.5 KiB | 00m00s [ 66/156] Installing p11-kit-0:0.25.5-3 100% | 275.9 MiB/s | 2.2 MiB | 00m00s [ 67/156] Installing libunistring-0:1.1 100% | 432.7 MiB/s | 1.7 MiB | 00m00s [ 68/156] Installing libidn2-0:2.3.7-2. 100% | 163.6 MiB/s | 335.1 KiB | 00m00s [ 69/156] Installing libpsl-0:0.21.5-4. 100% | 0.0 B/s | 81.7 KiB | 00m00s [ 70/156] Installing p11-kit-trust-0:0. 100% | 54.8 MiB/s | 393.1 KiB | 00m00s [ 71/156] Installing openssl-libs-1:3.2 100% | 366.0 MiB/s | 7.7 MiB | 00m00s [ 72/156] Installing coreutils-0:9.5-11 100% | 285.3 MiB/s | 5.7 MiB | 00m00s [ 73/156] Installing ca-certificates-0: 100% | 3.9 MiB/s | 2.4 MiB | 00m01s [ 74/156] Installing gzip-0:1.13-2.fc41 100% | 192.7 MiB/s | 394.6 KiB | 00m00s [ 75/156] Installing authselect-libs-0: 100% | 163.5 MiB/s | 837.2 KiB | 00m00s [ 76/156] Installing cracklib-0:2.9.11- 100% | 81.5 MiB/s | 250.3 KiB | 00m00s [ 77/156] Installing libpwquality-0:1.4 100% | 140.0 MiB/s | 430.1 KiB | 00m00s [ 78/156] Installing rpm-sequoia-0:1.7. 100% | 387.3 MiB/s | 2.3 MiB | 00m00s [ 79/156] Installing libevent-0:2.1.12- 100% | 292.8 MiB/s | 899.5 KiB | 00m00s [ 80/156] Installing zstd-0:1.5.6-2.fc4 100% | 422.9 MiB/s | 1.7 MiB | 00m00s [ 81/156] Installing util-linux-core-0: 100% | 247.6 MiB/s | 1.5 MiB | 00m00s [ 82/156] Installing tar-2:1.35-4.fc41. 100% | 369.8 MiB/s | 3.0 MiB | 00m00s [ 83/156] Installing libsemanage-0:3.7- 100% | 144.2 MiB/s | 295.2 KiB | 00m00s [ 84/156] Installing shadow-utils-2:4.1 100% | 166.6 MiB/s | 4.2 MiB | 00m00s [ 85/156] Installing libutempter-0:1.2. 100% | 58.3 MiB/s | 59.7 KiB | 00m00s [ 86/156] Installing zip-0:3.0-41.fc41. 100% | 230.2 MiB/s | 707.1 KiB | 00m00s [ 87/156] Installing gdbm-1:1.23-7.fc41 100% | 227.4 MiB/s | 465.8 KiB | 00m00s [ 88/156] Installing cyrus-sasl-lib-0:2 100% | 329.4 MiB/s | 2.3 MiB | 00m00s [ 89/156] Installing libfdisk-0:2.40.2- 100% | 355.5 MiB/s | 364.1 KiB | 00m00s [ 90/156] Installing bzip2-0:1.0.8-19.f 100% | 97.8 MiB/s | 100.2 KiB | 00m00s [ 91/156] Installing libxml2-0:2.12.9-1 100% | 337.2 MiB/s | 1.7 MiB | 00m00s [ 92/156] Installing libarchive-0:3.7.4 100% | 302.3 MiB/s | 928.6 KiB | 00m00s [ 93/156] Installing add-determinism-0: 100% | 392.6 MiB/s | 2.4 MiB | 00m00s [ 94/156] Installing build-reproducibil 100% | 0.0 B/s | 1.0 KiB | 00m00s [ 95/156] Installing sqlite-libs-0:3.46 100% | 357.3 MiB/s | 1.4 MiB | 00m00s [ 96/156] Installing rpm-libs-0:4.20.0- 100% | 355.2 MiB/s | 727.4 KiB | 00m00s [ 97/156] Installing ed-0:1.20.2-2.fc41 100% | 145.7 MiB/s | 149.2 KiB | 00m00s [ 98/156] Installing patch-0:2.7.6-25.f 100% | 261.9 MiB/s | 268.2 KiB | 00m00s [ 99/156] Installing elfutils-default-y 100% | 408.6 KiB/s | 2.0 KiB | 00m00s [100/156] Installing elfutils-libs-0:0. 100% | 213.8 MiB/s | 656.7 KiB | 00m00s [101/156] Installing cpio-0:2.15-2.fc41 100% | 274.9 MiB/s | 1.1 MiB | 00m00s [102/156] Installing diffutils-0:3.10-8 100% | 318.1 MiB/s | 1.6 MiB | 00m00s [103/156] Installing libpkgconf-0:2.3.0 100% | 0.0 B/s | 79.3 KiB | 00m00s [104/156] Installing pkgconf-0:2.3.0-1. 100% | 89.0 MiB/s | 91.1 KiB | 00m00s [105/156] Installing keyutils-libs-0:1. 100% | 0.0 B/s | 55.8 KiB | 00m00s [106/156] Installing libverto-0:0.3.2-9 100% | 30.5 MiB/s | 31.3 KiB | 00m00s [107/156] Installing krb5-libs-0:1.21.3 100% | 331.3 MiB/s | 2.3 MiB | 00m00s [108/156] Installing libtirpc-0:1.3.6-1 100% | 194.7 MiB/s | 199.4 KiB | 00m00s [109/156] Installing libnsl2-0:2.0.1-2. 100% | 57.7 MiB/s | 59.1 KiB | 00m00s [110/156] Installing pam-0:1.6.1-7.fc41 100% | 170.7 MiB/s | 1.9 MiB | 00m00s [111/156] Installing libssh-0:0.10.6-8. 100% | 251.7 MiB/s | 515.4 KiB | 00m00s [112/156] Installing jansson-0:2.13.1-1 100% | 87.6 MiB/s | 89.7 KiB | 00m00s [113/156] Installing libgomp-0:14.2.1-3 100% | 256.2 MiB/s | 524.8 KiB | 00m00s [114/156] Installing rpm-build-libs-0:4 100% | 202.6 MiB/s | 207.5 KiB | 00m00s [115/156] Installing json-c-0:0.17-4.fc 100% | 81.7 MiB/s | 83.6 KiB | 00m00s [116/156] Installing xxhash-libs-0:0.8. 100% | 0.0 B/s | 89.9 KiB | 00m00s [117/156] Installing libbrotli-0:1.1.0- 100% | 273.4 MiB/s | 839.9 KiB | 00m00s [118/156] Installing libnghttp2-0:1.62. 100% | 163.2 MiB/s | 167.1 KiB | 00m00s [119/156] Installing libtool-ltdl-0:2.4 100% | 65.7 MiB/s | 67.3 KiB | 00m00s [120/156] Installing openldap-0:2.6.8-6 100% | 208.1 MiB/s | 639.2 KiB | 00m00s [121/156] Installing libcurl-0:8.9.1-2. 100% | 400.0 MiB/s | 819.2 KiB | 00m00s [122/156] Installing elfutils-debuginfo 100% | 81.8 MiB/s | 83.8 KiB | 00m00s [123/156] Installing binutils-0:2.43.1- 100% | 375.2 MiB/s | 27.4 MiB | 00m00s [124/156] Installing elfutils-0:0.192-7 100% | 375.6 MiB/s | 2.6 MiB | 00m00s [125/156] Installing gdb-minimal-0:15.2 100% | 371.2 MiB/s | 13.0 MiB | 00m00s [126/156] Installing debugedit-0:5.1-2. 100% | 198.5 MiB/s | 203.3 KiB | 00m00s [127/156] Installing curl-0:8.9.1-2.fc4 100% | 70.9 MiB/s | 798.6 KiB | 00m00s [128/156] Installing rpm-0:4.20.0-1.fc4 100% | 179.0 MiB/s | 2.5 MiB | 00m00s [129/156] Installing efi-srpm-macros-0: 100% | 0.0 B/s | 41.2 KiB | 00m00s [130/156] Installing lua-srpm-macros-0: 100% | 0.0 B/s | 1.9 KiB | 00m00s [131/156] Installing zig-srpm-macros-0: 100% | 0.0 B/s | 1.7 KiB | 00m00s [132/156] Installing pkgconf-m4-0:2.3.0 100% | 0.0 B/s | 14.8 KiB | 00m00s [133/156] Installing pkgconf-pkg-config 100% | 0.0 B/s | 1.8 KiB | 00m00s [134/156] Installing rust-srpm-macros-0 100% | 0.0 B/s | 5.6 KiB | 00m00s [135/156] Installing qt5-srpm-macros-0: 100% | 0.0 B/s | 776.0 B | 00m00s [136/156] Installing perl-srpm-macros-0 100% | 0.0 B/s | 1.1 KiB | 00m00s [137/156] Installing package-notes-srpm 100% | 0.0 B/s | 2.0 KiB | 00m00s [138/156] Installing openblas-srpm-macr 100% | 0.0 B/s | 392.0 B | 00m00s [139/156] Installing ocaml-srpm-macros- 100% | 0.0 B/s | 2.2 KiB | 00m00s [140/156] Installing kernel-srpm-macros 100% | 0.0 B/s | 2.3 KiB | 00m00s [141/156] Installing gnat-srpm-macros-0 100% | 0.0 B/s | 1.3 KiB | 00m00s [142/156] Installing ghc-srpm-macros-0: 100% | 0.0 B/s | 1.0 KiB | 00m00s [143/156] Installing fpc-srpm-macros-0: 100% | 0.0 B/s | 420.0 B | 00m00s [144/156] Installing ansible-srpm-macro 100% | 0.0 B/s | 36.2 KiB | 00m00s [145/156] Installing qt6-srpm-macros-0: 100% | 0.0 B/s | 732.0 B | 00m00s [146/156] Installing forge-srpm-macros- 100% | 0.0 B/s | 40.3 KiB | 00m00s [147/156] Installing go-srpm-macros-0:3 100% | 0.0 B/s | 62.0 KiB | 00m00s [148/156] Installing fonts-srpm-macros- 100% | 0.0 B/s | 57.0 KiB | 00m00s [149/156] Installing python-srpm-macros 100% | 0.0 B/s | 52.2 KiB | 00m00s [150/156] Installing redhat-rpm-config- 100% | 185.6 MiB/s | 190.1 KiB | 00m00s [151/156] Installing rpm-build-0:4.20.0 100% | 99.0 MiB/s | 202.8 KiB | 00m00s [152/156] Installing pyproject-srpm-mac 100% | 2.4 MiB/s | 2.5 KiB | 00m00s [153/156] Installing util-linux-0:2.40. 100% | 170.4 MiB/s | 3.7 MiB | 00m00s [154/156] Installing authselect-0:1.5.0 100% | 79.1 MiB/s | 161.9 KiB | 00m00s [155/156] Installing which-0:2.21-42.fc 100% | 80.5 MiB/s | 82.4 KiB | 00m00s [156/156] Installing info-0:7.1-3.fc41. 100% | 436.4 KiB/s | 362.2 KiB | 00m01s Warning: skipped OpenPGP checks for 154 packages from repository: http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch Complete! Finish: installing minimal buildroot with dnf5 Start: creating root cache Finish: creating root cache Finish: chroot init INFO: Installed packages: INFO: add-determinism-0.3.6-3.fc41.x86_64 alternatives-1.31-1.fc41.x86_64 ansible-srpm-macros-1-16.fc41.noarch audit-libs-4.0.2-1.fc41.x86_64 authselect-1.5.0-8.fc41.x86_64 authselect-libs-1.5.0-8.fc41.x86_64 basesystem-11-21.fc41.noarch bash-5.2.32-1.fc41.x86_64 binutils-2.43.1-5.fc41.x86_64 build-reproducibility-srpm-macros-0.3.6-3.fc41.noarch bzip2-1.0.8-19.fc41.x86_64 bzip2-libs-1.0.8-19.fc41.x86_64 ca-certificates-2024.2.69_v8.0.401-1.0.fc41.noarch coreutils-9.5-11.fc41.x86_64 coreutils-common-9.5-11.fc41.x86_64 cpio-2.15-2.fc41.x86_64 cracklib-2.9.11-6.fc41.x86_64 crypto-policies-20241029-1.git8baf557.fc41.noarch curl-8.9.1-2.fc41.x86_64 cyrus-sasl-lib-2.1.28-27.fc41.x86_64 debugedit-5.1-2.fc41.x86_64 diffutils-3.10-8.fc41.x86_64 dwz-0.15-8.fc41.x86_64 ed-1.20.2-2.fc41.x86_64 efi-srpm-macros-5-13.fc41.noarch elfutils-0.192-7.fc41.x86_64 elfutils-debuginfod-client-0.192-7.fc41.x86_64 elfutils-default-yama-scope-0.192-7.fc41.noarch elfutils-libelf-0.192-7.fc41.x86_64 elfutils-libs-0.192-7.fc41.x86_64 fedora-gpg-keys-41-1.noarch fedora-release-41-29.noarch fedora-release-common-41-29.noarch fedora-release-identity-basic-41-29.noarch fedora-repos-41-1.noarch file-5.45-7.fc41.x86_64 file-libs-5.45-7.fc41.x86_64 filesystem-3.18-23.fc41.x86_64 findutils-4.10.0-4.fc41.x86_64 fonts-srpm-macros-2.0.5-17.fc41.noarch forge-srpm-macros-0.4.0-1.fc41.noarch fpc-srpm-macros-1.3-13.fc41.noarch gawk-5.3.0-4.fc41.x86_64 gdb-minimal-15.2-3.fc41.x86_64 gdbm-1.23-7.fc41.x86_64 gdbm-libs-1.23-7.fc41.x86_64 ghc-srpm-macros-1.9.1-2.fc41.noarch glibc-2.40-17.fc41.x86_64 glibc-common-2.40-17.fc41.x86_64 glibc-gconv-extra-2.40-17.fc41.x86_64 glibc-minimal-langpack-2.40-17.fc41.x86_64 gmp-6.3.0-2.fc41.x86_64 gnat-srpm-macros-6-6.fc41.noarch go-srpm-macros-3.6.0-5.fc41.noarch grep-3.11-9.fc41.x86_64 gzip-1.13-2.fc41.x86_64 info-7.1-3.fc41.x86_64 jansson-2.13.1-10.fc41.x86_64 json-c-0.17-4.fc41.x86_64 kernel-srpm-macros-1.0-24.fc41.noarch keyutils-libs-1.6.3-4.fc41.x86_64 krb5-libs-1.21.3-3.fc41.x86_64 libacl-2.3.2-2.fc41.x86_64 libarchive-3.7.4-4.fc41.x86_64 libattr-2.5.2-4.fc41.x86_64 libblkid-2.40.2-4.fc41.x86_64 libbrotli-1.1.0-5.fc41.x86_64 libcap-2.70-4.fc41.x86_64 libcap-ng-0.8.5-3.fc41.x86_64 libcom_err-1.47.1-6.fc41.x86_64 libcurl-8.9.1-2.fc41.x86_64 libeconf-0.6.2-3.fc41.x86_64 libevent-2.1.12-14.fc41.x86_64 libfdisk-2.40.2-4.fc41.x86_64 libffi-3.4.6-3.fc41.x86_64 libgcc-14.2.1-3.fc41.x86_64 libgomp-14.2.1-3.fc41.x86_64 libidn2-2.3.7-2.fc41.x86_64 libmount-2.40.2-4.fc41.x86_64 libnghttp2-1.62.1-2.fc41.x86_64 libnsl2-2.0.1-2.fc41.x86_64 libpkgconf-2.3.0-1.fc41.x86_64 libpsl-0.21.5-4.fc41.x86_64 libpwquality-1.4.5-11.fc41.x86_64 libselinux-3.7-5.fc41.x86_64 libsemanage-3.7-2.fc41.x86_64 libsepol-3.7-2.fc41.x86_64 libsmartcols-2.40.2-4.fc41.x86_64 libssh-0.10.6-8.fc41.x86_64 libssh-config-0.10.6-8.fc41.noarch libstdc++-14.2.1-3.fc41.x86_64 libtasn1-4.19.0-9.fc41.x86_64 libtirpc-1.3.6-1.rc3.fc41.x86_64 libtool-ltdl-2.4.7-12.fc41.x86_64 libunistring-1.1-8.fc41.x86_64 libutempter-1.2.1-15.fc41.x86_64 libuuid-2.40.2-4.fc41.x86_64 libverto-0.3.2-9.fc41.x86_64 libxcrypt-4.4.37-4.fc41.x86_64 libxml2-2.12.9-1.fc41.x86_64 libzstd-1.5.6-2.fc41.x86_64 lua-libs-5.4.6-6.fc41.x86_64 lua-srpm-macros-1-14.fc41.noarch lz4-libs-1.10.0-1.fc41.x86_64 mpfr-4.2.1-5.fc41.x86_64 ncurses-base-6.5-2.20240629.fc41.noarch ncurses-libs-6.5-2.20240629.fc41.x86_64 ocaml-srpm-macros-10-3.fc41.noarch openblas-srpm-macros-2-18.fc41.noarch openldap-2.6.8-6.fc41.x86_64 openssl-libs-3.2.2-10.fc41.x86_64 p11-kit-0.25.5-3.fc41.x86_64 p11-kit-trust-0.25.5-3.fc41.x86_64 package-notes-srpm-macros-0.5-12.fc41.noarch pam-1.6.1-7.fc41.x86_64 pam-libs-1.6.1-7.fc41.x86_64 patch-2.7.6-25.fc41.x86_64 pcre2-10.44-1.fc41.1.x86_64 pcre2-syntax-10.44-1.fc41.1.noarch perl-srpm-macros-1-56.fc41.noarch pkgconf-2.3.0-1.fc41.x86_64 pkgconf-m4-2.3.0-1.fc41.noarch pkgconf-pkg-config-2.3.0-1.fc41.x86_64 popt-1.19-7.fc41.x86_64 publicsuffix-list-dafsa-20240107-4.fc41.noarch pyproject-srpm-macros-1.16.3-1.fc41.noarch python-srpm-macros-3.13-3.fc41.noarch qt5-srpm-macros-5.15.15-1.fc41.noarch qt6-srpm-macros-6.8.1-4.fc41.noarch readline-8.2-10.fc41.x86_64 redhat-rpm-config-293-1.fc41.noarch rpm-4.20.0-1.fc41.x86_64 rpm-build-4.20.0-1.fc41.x86_64 rpm-build-libs-4.20.0-1.fc41.x86_64 rpm-libs-4.20.0-1.fc41.x86_64 rpm-sequoia-1.7.0-3.fc41.x86_64 rust-srpm-macros-26.3-3.fc41.noarch sed-4.9-3.fc41.x86_64 setup-2.15.0-8.fc41.noarch shadow-utils-4.15.1-12.fc41.x86_64 sqlite-libs-3.46.1-1.fc41.x86_64 systemd-libs-256.10-1.fc41.x86_64 tar-1.35-4.fc41.x86_64 unzip-6.0-64.fc41.x86_64 util-linux-2.40.2-4.fc41.x86_64 util-linux-core-2.40.2-4.fc41.x86_64 which-2.21-42.fc41.x86_64 xxhash-libs-0.8.3-1.fc41.x86_64 xz-5.6.2-2.fc41.x86_64 xz-libs-5.6.2-2.fc41.x86_64 zig-srpm-macros-1-3.fc41.noarch zip-3.0-41.fc41.x86_64 zlib-ng-compat-2.1.7-3.fc41.x86_64 zstd-1.5.6-2.fc41.x86_64 Start: buildsrpm Start: rpmbuild -bs Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1725321600 Wrote: /builddir/build/SRPMS/python-torch-2.4.0-10.fc41.src.rpm Finish: rpmbuild -bs INFO: chroot_scan: 1 files copied to /var/lib/copr-rpmbuild/results/chroot_scan INFO: /var/lib/mock/fedora-41-x86_64-1736346719.362177/root/var/log/dnf5.log INFO: chroot_scan: creating tarball /var/lib/copr-rpmbuild/results/chroot_scan.tar.gz /bin/tar: Removing leading `/' from member names Finish: buildsrpm INFO: Done(/var/lib/copr-rpmbuild/workspace/workdir-3eg6jxui/python-torch/python-torch.spec) Config(child) 0 minutes 17 seconds INFO: Results and/or logs in: /var/lib/copr-rpmbuild/results INFO: Cleaning up build root ('cleanup_on_success=True') Start: clean chroot INFO: unmounting tmpfs. Finish: clean chroot INFO: Start(/var/lib/copr-rpmbuild/results/python-torch-2.4.0-10.fc41.src.rpm) Config(fedora-41-x86_64) Start(bootstrap): chroot init INFO: mounting tmpfs at /var/lib/mock/fedora-41-x86_64-bootstrap-1736346719.362177/root. INFO: reusing tmpfs at /var/lib/mock/fedora-41-x86_64-bootstrap-1736346719.362177/root. INFO: calling preinit hooks INFO: enabled root cache INFO: enabled package manager cache Start(bootstrap): cleaning package manager metadata Finish(bootstrap): cleaning package manager metadata Finish(bootstrap): chroot init Start: chroot init INFO: mounting tmpfs at /var/lib/mock/fedora-41-x86_64-1736346719.362177/root. INFO: calling preinit hooks INFO: enabled root cache Start: unpacking root cache Finish: unpacking root cache INFO: enabled package manager cache Start: cleaning package manager metadata Finish: cleaning package manager metadata INFO: enabled HW Info plugin INFO: Buildroot is handled by package management downloaded with a bootstrap image: rpm-4.20.0-1.fc41.x86_64 rpm-sequoia-1.7.0-3.fc41.x86_64 dnf5-5.2.8.1-2.fc41.x86_64 dnf5-plugins-5.2.8.1-2.fc41.x86_64 Finish: chroot init Start: build phase for python-torch-2.4.0-10.fc41.src.rpm Start: build setup for python-torch-2.4.0-10.fc41.src.rpm Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1725321600 Wrote: /builddir/build/SRPMS/python-torch-2.4.0-10.fc41.src.rpm Updating and loading repositories: updates 100% | 211.0 KiB/s | 22.4 KiB | 00m00s fedora 100% | 391.2 KiB/s | 29.7 KiB | 00m00s Copr repository 100% | 89.1 KiB/s | 1.5 KiB | 00m00s Additional repo http_kojipkgs_fedorapr 100% | 191.6 KiB/s | 3.8 KiB | 00m00s Repositories loaded. Package Arch Version Repository Size Installing: FP16-devel noarch 1.0^git20210317.0a92994-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 32.7 KiB cmake x86_64 3.30.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 32.5 MiB cpuinfo-devel x86_64 23.11.04-0.gitd6860c4.fc41.1 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 65.2 KiB eigen3-devel noarch 3.4.0-16.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.4 MiB flexiblas-devel x86_64 3.4.4-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.8 MiB fmt-devel x86_64 11.0.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 575.3 KiB foxi-devel x86_64 1.4.1^git20210526.c278588-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 119.6 KiB fxdiv-devel noarch 1.0^git20201208.63058ef-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.9 KiB gcc-c++ x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 38.2 MiB gcc-gfortran x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 37.2 MiB hipblas-devel x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.8 MiB hipblaslt-devel x86_64 6.2.0-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 166.4 KiB hipcub-devel x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.1 MiB hipfft-devel x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 140.8 KiB hiprand-devel x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 287.8 KiB hipsolver-devel x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 689.2 KiB hipsparse-devel x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 799.9 KiB libomp-devel x86_64 19.1.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 31.2 MiB magma-devel x86_64 2.8.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.6 MiB miopen-devel x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 548.3 MiB ninja-build x86_64 1.12.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 432.6 KiB numactl-devel x86_64 2.0.19-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 26.7 KiB onnx-devel x86_64 1.15.0-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.1 MiB openmpi-devel x86_64 5.0.5-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.2 MiB pocketfft-devel noarch 1.0^git20230214.076cb3d-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 120.0 KiB protobuf-devel x86_64 3.19.6-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.7 MiB psimd-devel noarch 20200517.072586a-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 45.6 KiB pthreadpool-devel x86_64 0.0^git20230829.4fe0e1e-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 99.1 KiB python3-devel x86_64 3.13.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 MiB python3-filelock noarch 3.15.4-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 90.2 KiB python3-fsspec noarch 2024.12.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB python3-jinja2 noarch 3.1.5-1.fc41 copr_base 2.9 MiB python3-networkx noarch 3.3-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.8 MiB python3-numpy x86_64 1:1.26.4-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 43.8 MiB python3-pybind11 x86_64 2.13.6-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 897.4 KiB python3-pyyaml x86_64 6.0.1-18.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 791.1 KiB python3-setuptools noarch 69.2.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 7.2 MiB python3-sphinx noarch 1:7.3.7-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.8 MiB python3-sympy x86_64 1.13.3-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 78.3 MiB python3-typing-extensions noarch 4.12.2-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 464.5 KiB rccl-devel x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 131.1 KiB rocblas-devel x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.5 MiB rocfft-devel x86_64 6.2.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 117.2 KiB rocm-cmake noarch 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 129.2 KiB rocm-comgr-devel x86_64 18-10.rocm6.2.1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 103.1 KiB rocm-core-devel x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.1 KiB rocm-hip-devel x86_64 6.2.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.6 MiB rocm-rpm-macros x86_64 6.2-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 19.1 KiB rocm-rpm-macros-modules x86_64 6.2-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 24.2 KiB rocm-runtime-devel x86_64 6.2.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 556.1 KiB rocprim-devel noarch 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.6 MiB rocrand-devel x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.8 MiB rocthrust-devel x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.3 MiB roctracer-devel x86_64 6.2.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 424.0 KiB sleef-devel x86_64 3.7.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 273.1 KiB valgrind-devel x86_64 1:3.24.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 501.8 KiB xnnpack-devel x86_64 0.0^git20240229.fcbf55a-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 225.1 KiB Installing dependencies: annobin-docs noarch 12.69-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 97.7 KiB annobin-plugin-gcc x86_64 12.69-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 985.0 KiB blas x86_64 3.12.0-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 MiB boost-atomic x86_64 1.83.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 21.0 KiB boost-filesystem x86_64 1.83.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 147.6 KiB boost-system x86_64 1.83.0-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.3 KiB clang-resource-filesystem x86_64 19.1.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 15.3 KiB clang18 x86_64 18.1.8-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 644.4 KiB clang18-devel x86_64 18.1.8-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 23.7 MiB clang18-libs x86_64 18.1.8-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 102.1 MiB clang18-resource-filesystem x86_64 18.1.8-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B clang18-tools-extra x86_64 18.1.8-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 85.3 MiB cmake-data noarch 3.30.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.2 MiB cmake-filesystem x86_64 3.30.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B cmake-rpm-macros noarch 3.30.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 7.5 KiB compiler-rt18 x86_64 18.1.8-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 28.0 MiB cpp x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 35.0 MiB cpuinfo x86_64 23.11.04-0.gitd6860c4.fc41.1 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 114.3 KiB dbus x86_64 1:1.14.10-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B dbus-broker x86_64 36-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 382.8 KiB dbus-common noarch 1:1.14.10-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 11.2 KiB emacs-filesystem noarch 1:30.0-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 0.0 B environment-modules x86_64 5.4.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB expat x86_64 2.6.4-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 292.9 KiB flexiblas x86_64 3.4.4-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 48.5 KiB flexiblas-netlib x86_64 3.4.4-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.7 MiB flexiblas-netlib64 x86_64 3.4.4-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.8 MiB flexiblas-openblas-openmp x86_64 3.4.4-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 43.3 KiB flexiblas-openblas-openmp64 x86_64 3.4.4-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 43.3 KiB fmt x86_64 11.0.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 264.1 KiB foxi x86_64 1.4.1^git20210526.c278588-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.1 KiB gcc x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 104.3 MiB gcc-plugin-annobin x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 61.1 KiB glibc-devel x86_64 2.40-17.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.3 MiB groff-base x86_64 1.23.0-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.8 MiB hipblas x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.9 MiB hipblas-gfx1100 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 995.3 KiB hipblas-gfx1103 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 995.3 KiB hipblas-gfx90a x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 995.3 KiB hipblas-gfx942 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 995.3 KiB hipblaslt x86_64 6.2.0-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 882.8 MiB hipcc x86_64 18-10.rocm6.2.1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 655.9 KiB hipfft x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 577.7 KiB hiprand x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 97.3 KiB hiprand-gfx1100 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 23.3 KiB hiprand-gfx1103 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 23.3 KiB hiprand-gfx90a x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 23.3 KiB hiprand-gfx942 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 23.3 KiB hipsolver x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.3 MiB hipsolver-gfx1100 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 325.4 KiB hipsolver-gfx1103 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 325.4 KiB hipsolver-gfx90a x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 325.4 KiB hipsolver-gfx942 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 325.4 KiB hipsparse x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 981.2 KiB hipsparse-gfx1100 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 245.0 KiB hipsparse-gfx1103 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 245.1 KiB hipsparse-gfx90a x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 245.0 KiB hipsparse-gfx942 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 245.0 KiB hsakmt x86_64 1.0.6-46.rocm6.2.1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 181.5 KiB hsakmt-devel x86_64 1.0.6-46.rocm6.2.1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 110.6 KiB hwdata noarch 0.391-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 9.4 MiB hwloc-libs x86_64 2.11.2-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.9 MiB jsoncpp x86_64 1.9.5-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 253.4 KiB kernel-headers x86_64 6.12.4-200.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 6.4 MiB kmod x86_64 33-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 235.5 KiB lapack x86_64 3.12.0-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.7 MiB less x86_64 661-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 405.3 KiB libatomic x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 40.7 KiB libb2 x86_64 0.98.1-12.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 42.2 KiB libcbor x86_64 0.11.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 73.9 KiB libdrm x86_64 2.4.124-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 397.7 KiB libedit x86_64 3.1-54.20250104cvs.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 239.3 KiB libedit-devel x86_64 3.1-54.20250104cvs.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 59.4 KiB libfabric x86_64 1.22.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.2 MiB libfido2 x86_64 1.15.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 238.2 KiB libgfortran x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.0 MiB libibverbs x86_64 51.0-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.2 MiB libmpc x86_64 1.3.1-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 164.7 KiB libnl3 x86_64 3.11.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 MiB libomp x86_64 19.1.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 49.8 MiB libpciaccess x86_64 0.16-13.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 44.6 KiB libpipeline x86_64 1.5.7-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 124.3 KiB libpsm2 x86_64 11.2.230-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 442.5 KiB libquadmath x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 325.9 KiB libquadmath-devel x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 21.9 KiB librdmacm x86_64 51.0-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 154.2 KiB libseccomp x86_64 2.5.5-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 173.3 KiB libstdc++-devel x86_64 14.2.1-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 15.4 MiB libuv x86_64 1:1.49.2-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 565.0 KiB libxcrypt-devel x86_64 4.4.37-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 30.5 KiB libyaml x86_64 0.2.5-15.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 134.4 KiB lld18 x86_64 18.1.8-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 134.5 KiB lld18-devel x86_64 18.1.8-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 38.7 KiB lld18-libs x86_64 18.1.8-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.3 MiB llvm-libs x86_64 19.1.5-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 123.0 MiB llvm18 x86_64 18.1.8-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 112.2 MiB llvm18-devel x86_64 18.1.8-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 24.2 MiB llvm18-googletest x86_64 18.1.8-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.2 MiB llvm18-libs x86_64 18.1.8-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 113.5 MiB llvm18-static x86_64 18.1.8-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 283.9 MiB llvm18-test x86_64 18.1.8-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.9 MiB logrotate x86_64 3.22.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 153.1 KiB magma x86_64 2.8.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.3 GiB make x86_64 1:4.4.1-8.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 MiB man-db x86_64 2.12.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.8 MiB miopen x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.4 GiB miopen-gfx1100 x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 356.0 MiB miopen-gfx1103 x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 356.0 MiB miopen-gfx90a x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 356.0 MiB miopen-gfx942 x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 356.0 MiB mpdecimal x86_64 2.5.1-16.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 204.9 KiB munge x86_64 0.5.16-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 346.3 KiB munge-libs x86_64 0.5.16-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 32.1 KiB ncurses x86_64 6.5-2.20240629.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 627.3 KiB ncurses-c++-libs x86_64 6.5-2.20240629.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 161.7 KiB ncurses-devel x86_64 6.5-2.20240629.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 870.1 KiB numactl-libs x86_64 2.0.19-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 61.0 KiB onnx-libs x86_64 1.15.0-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.2 MiB openblas x86_64 0.3.26-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 96.0 KiB openblas-openmp x86_64 0.3.26-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 39.4 MiB openblas-openmp64 x86_64 0.3.26-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 39.5 MiB openmpi x86_64 5.0.5-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 7.0 MiB openssh x86_64 9.8p1-3.fc41.2 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.8 MiB openssh-clients x86_64 9.8p1-3.fc41.2 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.6 MiB orangefs x86_64 2.9.8-12.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.1 MiB perl-AutoLoader noarch 5.74-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 20.5 KiB perl-B x86_64 1.89-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 498.0 KiB perl-Carp noarch 1.54-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 46.6 KiB perl-Class-Struct noarch 0.68-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 25.4 KiB perl-Data-Dumper x86_64 2.189-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 111.7 KiB perl-Digest noarch 1.20-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 35.3 KiB perl-Digest-MD5 x86_64 2.59-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 59.8 KiB perl-DynaLoader x86_64 1.56-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 32.1 KiB perl-Encode x86_64 4:3.21-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.7 MiB perl-Errno x86_64 1.38-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.4 KiB perl-Exporter noarch 5.78-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 54.3 KiB perl-Fcntl x86_64 1.18-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 49.0 KiB perl-File-Basename noarch 2.86-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 14.0 KiB perl-File-Copy noarch 2.41-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 19.6 KiB perl-File-Path noarch 2.18-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 63.5 KiB perl-File-Temp noarch 1:0.231.100-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 162.3 KiB perl-File-Which noarch 1.27-12.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 30.4 KiB perl-File-stat noarch 1.14-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 12.5 KiB perl-FileHandle noarch 2.05-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 9.3 KiB perl-Getopt-Long noarch 1:2.58-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 144.5 KiB perl-Getopt-Std noarch 1.14-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 11.2 KiB perl-HTTP-Tiny noarch 0.090-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 154.4 KiB perl-IO x86_64 1.55-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 151.1 KiB perl-IO-Socket-IP noarch 0.43-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 100.3 KiB perl-IO-Socket-SSL noarch 2.089-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 703.3 KiB perl-IPC-Open3 noarch 1.22-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 22.5 KiB perl-MIME-Base32 noarch 1.303-21.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 30.7 KiB perl-MIME-Base64 x86_64 3.16-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 46.1 KiB perl-Net-SSLeay x86_64 1.94-7.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.3 MiB perl-POSIX x86_64 2.20-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 235.1 KiB perl-PathTools x86_64 3.91-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 180.0 KiB perl-Pod-Escapes noarch 1:1.07-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 24.9 KiB perl-Pod-Perldoc noarch 3.28.01-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 163.7 KiB perl-Pod-Simple noarch 1:3.45-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 560.9 KiB perl-Pod-Usage noarch 4:2.03-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 84.8 KiB perl-Scalar-List-Utils x86_64 5:1.68-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 148.9 KiB perl-SelectSaver noarch 1.02-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.2 KiB perl-Socket x86_64 4:2.038-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 124.0 KiB perl-Storable x86_64 1:3.32-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 232.4 KiB perl-Symbol noarch 1.09-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 6.8 KiB perl-Term-ANSIColor noarch 5.01-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 97.5 KiB perl-Term-Cap noarch 1.18-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 29.3 KiB perl-Text-ParseWords noarch 3.31-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 13.6 KiB perl-Text-Tabs+Wrap noarch 2024.001-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 22.6 KiB perl-Time-Local noarch 2:1.350-511.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 69.0 KiB perl-URI noarch 5.30-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 256.9 KiB perl-base noarch 2.27-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 12.5 KiB perl-constant noarch 1.33-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 26.2 KiB perl-if noarch 0.61.000-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.8 KiB perl-interpreter x86_64 4:5.40.0-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 122.3 KiB perl-lib x86_64 0.65-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.5 KiB perl-libnet noarch 3.15-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 289.4 KiB perl-libs x86_64 4:5.40.0-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 9.9 MiB perl-locale noarch 1.12-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 6.5 KiB perl-mro x86_64 1.29-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 45.6 KiB perl-overload noarch 1.37-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 71.5 KiB perl-overloading noarch 0.02-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.8 KiB perl-parent noarch 1:0.242-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.0 KiB perl-podlators noarch 1:6.0.2-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 317.5 KiB perl-vars noarch 1.05-512.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.9 KiB pmix x86_64 4.2.8-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.0 MiB procps-ng x86_64 4.0.4-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 MiB protobuf x86_64 3.19.6-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.3 MiB protobuf-compiler x86_64 3.19.6-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.5 MiB prrte x86_64 3.0.6-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 175.0 KiB prrte-libs x86_64 3.0.6-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB pthreadpool x86_64 0.0^git20230829.4fe0e1e-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 113.3 KiB pybind11-devel x86_64 2.13.6-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 847.8 KiB pyproject-rpm-macros noarch 1.16.3-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 113.7 KiB python-pip-wheel noarch 24.2-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.2 MiB python-rpm-macros noarch 3.13-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 22.1 KiB python3 x86_64 3.13.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 30.6 KiB python3-babel noarch 2.16.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 28.3 MiB python3-charset-normalizer noarch 3.3.2-5.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 314.8 KiB python3-docutils noarch 0.20.1-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.8 MiB python3-idna noarch 3.7-4.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 595.8 KiB python3-imagesize noarch 1.4.1-9.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 35.3 KiB python3-libs x86_64 3.13.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 40.2 MiB python3-markupsafe x86_64 2.1.5-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 57.5 KiB python3-mpmath noarch 1.3.0-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.2 MiB python3-packaging noarch 24.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 422.3 KiB python3-peachpy noarch 0.2.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 13.3 MiB python3-pygments noarch 2.18.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.6 MiB python3-requests noarch 2.32.3-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 485.9 KiB python3-rpm-generators noarch 14-11.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 81.7 KiB python3-rpm-macros noarch 3.13-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 6.4 KiB python3-six noarch 1.16.0-23.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 118.3 KiB python3-snowballstemmer noarch 2.2.0-13.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.7 MiB python3-sphinx-theme-alabaster noarch 0.7.16-6.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 41.9 KiB python3-urllib3 noarch 1.26.20-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 MiB rccl x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.1 GiB rccl-data noarch 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.9 MiB rhash x86_64 1.4.4-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 349.9 KiB rocblas x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.0 GiB rocfft x86_64 6.2.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 54.3 MiB rocm-comgr x86_64 18-10.rocm6.2.1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.9 MiB rocm-core x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.4 KiB rocm-device-libs x86_64 18-10.rocm6.2.1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 3.2 MiB rocm-hip x86_64 6.2.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 22.9 MiB rocm-runtime x86_64 6.2.1-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.7 MiB rocm-smi x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.4 MiB rocminfo x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 81.6 KiB rocrand x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 422.0 MiB rocrand-gfx1100 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 44.8 MiB rocrand-gfx1103 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 44.8 MiB rocrand-gfx90a x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 57.3 MiB rocrand-gfx942 x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 43.2 MiB rocsolver x86_64 6.2.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.3 GiB rocsparse x86_64 6.2.1-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 5.2 GiB roctracer x86_64 6.2.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.0 MiB rpm-mpi-hooks noarch 8-10.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 8.7 KiB sleef x86_64 3.7.0-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.2 MiB suitesparse x86_64 7.7.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 139.9 MiB systemd x86_64 256.10-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 16.7 MiB systemd-pam x86_64 256.10-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.1 MiB systemd-rpm-macros noarch 256.10-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 10.7 KiB tcl x86_64 1:8.6.14-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 4.2 MiB tcsh x86_64 6.24.13-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.2 MiB torque-libs x86_64 6.1.3-13.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 454.1 KiB tzdata noarch 2024b-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.6 MiB ucx x86_64 1.17.0-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 2.4 MiB valgrind x86_64 1:3.24.0-2.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 30.1 MiB vim-filesystem noarch 2:9.1.984-1.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 40.0 B xnnpack x86_64 0.0^git20240229.fcbf55a-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 1.9 MiB zlib-ng-compat-devel x86_64 2.1.7-3.fc41 http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch 106.8 KiB Transaction Summary: Installing: 307 packages Total size of inbound packages is 2 GiB. Need to download 2 GiB. After this operation, 28 GiB extra will be used (install 28 GiB, remove 0 B). [ 1/29] python3-devel-0:3.13.1-2.fc41.x 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 2/31] python3-filelock-0:3.15.4-2.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 3/33] python3-jinja2-0:3.1.5-1.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 4/36] python3-pyyaml-0:6.0.1-18.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 5/37] python3-setuptools-0:69.2.0-8.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 6/38] python3-sphinx-1:7.3.7-2.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 7/62] perl-File-Basename-0:2.86-512.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 8/63] perl-interpreter-4:5.40.0-512.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 9/64] perl-libs-4:5.40.0-512.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [10/65] python3-0:3.13.1-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [11/83] perl-Getopt-Std-0:1.14-512.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [12/84] perl-Scalar-List-Utils-5:1.68-1 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [13/86] perl-PathTools-0:3.91-511.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [14/87] perl-URI-0:5.30-1.fc41.noarch 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 15/102] python3-babel-0:2.16.0-1.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 16/103] python3-docutils-0:0.20.1-6.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 17/104] python3-imagesize-0:1.4.1-9.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 18/105] python3-packaging-0:24.1-2.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 19/106] python3-pygments-0:2.18.0-2.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 20/107] python3-requests-0:2.32.3-3.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 21/108] python3-snowballstemmer-0:2.2 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 22/109] python3-sphinx-theme-alabaste 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 23/110] python3-charset-normalizer-0: 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 24/111] python3-idna-0:3.7-4.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 25/112] python3-urllib3-0:1.26.20-1.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 26/113] libyaml-0:0.2.5-15.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 27/116] python3-libs-0:3.13.1-2.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 28/120] perl-DynaLoader-0:1.56-512.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 29/121] perl-Carp-0:1.54-511.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 30/122] perl-Encode-4:3.21-511.fc41.x 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 31/123] perl-Exporter-0:5.78-511.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 32/124] perl-base-0:2.27-512.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 33/125] perl-overload-0:1.37-512.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 34/126] perl-Data-Dumper-0:2.189-512. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 35/127] perl-MIME-Base32-0:1.303-21.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 36/128] perl-MIME-Base64-0:3.16-511.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 37/129] perl-constant-0:1.33-512.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 38/130] perl-libnet-0:3.15-512.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 39/131] perl-parent-1:0.242-1.fc41.no 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 40/132] perl-Errno-0:1.38-512.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 41/133] perl-vars-0:1.05-512.fc41.noa 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 42/134] perl-Getopt-Long-1:2.58-2.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 43/135] perl-Storable-1:3.32-511.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 44/136] perl-B-0:1.89-512.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 45/137] perl-if-0:0.61.000-512.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 46/138] perl-overloading-0:0.02-512.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 47/142] openssh-clients-0:9.8p1-3.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 48/162] libedit-0:3.1-54.20250104cvs. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 49/198] perl-File-Temp-1:0.231.100-51 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 50/217] less-0:661-2.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 51/223] expat-0:2.6.4-1.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 52/232] python3-markupsafe-0:2.1.5-3. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 53/235] python3-six-0:1.16.0-23.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 54/236] tzdata-0:2024b-1.fc41.noarch 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 55/237] libb2-0:0.98.1-12.fc41.x86_64 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 56/238] mpdecimal-0:2.5.1-16.fc41.x86 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 57/239] python-pip-wheel-0:24.2-1.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 58/242] perl-mro-0:1.29-512.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 59/243] perl-Fcntl-0:1.18-512.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 60/244] perl-FileHandle-0:2.05-512.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 61/245] perl-IO-0:1.55-512.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 62/246] perl-IO-Socket-IP-0:0.43-1.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 63/247] perl-POSIX-0:2.20-512.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 64/248] perl-Symbol-0:1.09-512.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 65/249] perl-Digest-MD5-0:2.59-5.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 66/250] perl-Socket-4:2.038-511.fc41. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 67/251] perl-Time-Local-2:1.350-511.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 68/252] perl-locale-0:1.12-512.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 69/253] perl-File-stat-0:1.14-512.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 70/254] perl-SelectSaver-0:1.02-512.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 71/255] perl-Pod-Usage-4:2.03-511.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 72/256] perl-Text-ParseWords-0:3.31-5 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 73/257] perl-Class-Struct-0:0.68-512. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 74/258] perl-File-Path-0:2.18-511.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 75/259] perl-Digest-0:1.20-511.fc41.n 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 76/261] openssh-0:9.8p1-3.fc41.2.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 77/262] libfido2-0:1.15.0-2.fc41.x86_ 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 78/264] groff-base-0:1.23.0-7.fc41.x8 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 79/268] libcbor-0:0.11.0-2.fc41.x86_6 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 80/274] perl-Pod-Perldoc-0:3.28.01-51 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 81/275] perl-podlators-1:6.0.2-2.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 82/276] perl-HTTP-Tiny-0:0.090-1.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 83/277] perl-IPC-Open3-0:1.22-512.fc4 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 84/278] perl-Pod-Simple-1:3.45-511.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 85/279] perl-IO-Socket-SSL-0:2.089-1. 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 86/280] perl-Net-SSLeay-0:1.94-7.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 87/284] perl-Term-ANSIColor-0:5.01-51 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 88/285] perl-Term-Cap-0:1.18-511.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 89/286] ncurses-0:6.5-2.20240629.fc41 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 90/287] perl-Pod-Escapes-1:1.07-511.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 91/288] perl-Text-Tabs+Wrap-0:2024.00 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 92/289] perl-AutoLoader-0:5.74-512.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 93/300] pyproject-rpm-macros-0:1.16.3 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 94/301] python-rpm-macros-0:3.13-3.fc 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 95/302] python3-rpm-generators-0:14-1 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 96/303] python3-rpm-macros-0:3.13-3.f 100% | 0.0 B/s | 0.0 B | 00m00s >>> Already downloaded [ 97/307] FP16-devel-0:1.0^git20210317. 100% | 802.6 KiB/s | 17.7 KiB | 00m00s [ 98/307] eigen3-devel-0:3.4.0-16.fc41. 100% | 57.6 MiB/s | 1.1 MiB | 00m00s [ 99/307] cpuinfo-devel-0:23.11.04-0.gi 100% | 540.9 KiB/s | 22.7 KiB | 00m00s [100/307] fmt-devel-0:11.0.2-2.fc41.x86 100% | 15.8 MiB/s | 129.6 KiB | 00m00s [101/307] flexiblas-devel-0:3.4.4-3.fc4 100% | 11.1 MiB/s | 113.2 KiB | 00m00s [102/307] foxi-devel-0:1.4.1^git2021052 100% | 2.9 MiB/s | 23.5 KiB | 00m00s [103/307] fxdiv-devel-0:1.0^git20201208 100% | 1.5 MiB/s | 12.6 KiB | 00m00s [104/307] gcc-c++-0:14.2.1-3.fc41.x86_6 100% | 191.4 MiB/s | 14.2 MiB | 00m00s [105/307] gcc-gfortran-0:14.2.1-3.fc41. 100% | 145.7 MiB/s | 13.4 MiB | 00m00s [106/307] cmake-0:3.30.5-1.fc41.x86_64 100% | 63.9 MiB/s | 10.7 MiB | 00m00s [107/307] hipblas-devel-0:6.2.0-1.fc41. 100% | 2.8 MiB/s | 94.5 KiB | 00m00s [108/307] hipblaslt-devel-0:6.2.0-3.fc4 100% | 2.0 MiB/s | 34.5 KiB | 00m00s [109/307] hipfft-devel-0:6.2.1-1.fc41.x 100% | 1.2 MiB/s | 26.1 KiB | 00m00s [110/307] hipcub-devel-0:6.2.1-1.fc41.x 100% | 3.2 MiB/s | 99.6 KiB | 00m00s [111/307] hiprand-devel-0:6.2.0-1.fc41. 100% | 1.2 MiB/s | 38.0 KiB | 00m00s [112/307] libomp-devel-0:19.1.5-1.fc41. 100% | 63.1 MiB/s | 646.3 KiB | 00m00s [113/307] hipsparse-devel-0:6.2.0-1.fc4 100% | 2.3 MiB/s | 56.5 KiB | 00m00s [114/307] hipsolver-devel-0:6.2.0-1.fc4 100% | 1.2 MiB/s | 38.4 KiB | 00m00s [115/307] ninja-build-0:1.12.1-3.fc41.x 100% | 25.1 MiB/s | 180.0 KiB | 00m00s [116/307] magma-devel-0:2.8.0-2.fc41.x8 100% | 2.9 MiB/s | 105.6 KiB | 00m00s [117/307] numactl-devel-0:2.0.19-1.fc41 100% | 1.1 MiB/s | 22.6 KiB | 00m00s [118/307] openmpi-devel-0:5.0.5-2.fc41. 100% | 122.1 MiB/s | 2.2 MiB | 00m00s [119/307] onnx-devel-0:1.15.0-4.fc41.x8 100% | 4.8 MiB/s | 137.5 KiB | 00m00s [120/307] pocketfft-devel-0:1.0^git2023 100% | 4.5 MiB/s | 32.4 KiB | 00m00s [121/307] protobuf-devel-0:3.19.6-10.fc 100% | 35.9 MiB/s | 368.1 KiB | 00m00s [122/307] psimd-devel-0:20200517.072586 100% | 1.6 MiB/s | 12.8 KiB | 00m00s [123/307] pthreadpool-devel-0:0.0^git20 100% | 1.7 MiB/s | 14.3 KiB | 00m00s [124/307] python3-pybind11-0:2.13.6-1.f 100% | 21.3 MiB/s | 196.6 KiB | 00m00s [125/307] python3-fsspec-0:2024.12.0-1. 100% | 30.3 MiB/s | 403.3 KiB | 00m00s [126/307] python3-networkx-0:3.3-1.fc41 100% | 146.2 MiB/s | 3.1 MiB | 00m00s [127/307] python3-numpy-1:1.26.4-8.fc41 100% | 180.2 MiB/s | 7.4 MiB | 00m00s [128/307] python3-typing-extensions-0:4 100% | 15.3 MiB/s | 93.8 KiB | 00m00s [129/307] python3-sympy-0:1.13.3-1.fc41 100% | 234.1 MiB/s | 14.7 MiB | 00m00s [130/307] rccl-devel-0:6.2.1-1.fc41.x86 100% | 1.1 MiB/s | 28.0 KiB | 00m00s [131/307] rocfft-devel-0:6.2.1-2.fc41.x 100% | 1.1 MiB/s | 24.3 KiB | 00m00s [132/307] rocblas-devel-0:6.2.1-1.fc41. 100% | 3.5 MiB/s | 100.0 KiB | 00m00s [133/307] rocm-comgr-devel-0:18-10.rocm 100% | 4.1 MiB/s | 29.2 KiB | 00m00s [134/307] rocm-cmake-0:6.2.0-1.fc41.noa 100% | 1.6 MiB/s | 37.4 KiB | 00m00s [135/307] rocm-core-devel-0:6.2.0-1.fc4 100% | 584.4 KiB/s | 9.3 KiB | 00m00s [136/307] rocm-hip-devel-0:6.2.1-2.fc41 100% | 18.2 MiB/s | 223.0 KiB | 00m00s [137/307] rocm-rpm-macros-0:6.2-1.fc41. 100% | 1.8 MiB/s | 16.8 KiB | 00m00s [138/307] rocm-rpm-macros-modules-0:6.2 100% | 4.0 MiB/s | 20.6 KiB | 00m00s [139/307] rocm-runtime-devel-0:6.2.1-2. 100% | 9.0 MiB/s | 92.4 KiB | 00m00s [140/307] rocprim-devel-0:6.2.1-1.fc41. 100% | 9.2 MiB/s | 234.7 KiB | 00m00s [141/307] rocrand-devel-0:6.2.0-1.fc41. 100% | 8.5 MiB/s | 573.5 KiB | 00m00s [142/307] miopen-devel-0:6.2.1-1.fc41.x 100% | 89.0 MiB/s | 28.9 MiB | 00m00s [143/307] rocthrust-devel-0:6.2.1-1.fc4 100% | 6.8 MiB/s | 518.5 KiB | 00m00s [144/307] valgrind-devel-1:3.24.0-2.fc4 100% | 8.0 MiB/s | 49.2 KiB | 00m00s [145/307] sleef-devel-0:3.7.0-1.fc41.x8 100% | 1.2 MiB/s | 26.2 KiB | 00m00s [146/307] roctracer-devel-0:6.2.0-2.fc4 100% | 662.6 KiB/s | 37.8 KiB | 00m00s [147/307] xnnpack-devel-0:0.0^git202402 100% | 1.1 MiB/s | 26.8 KiB | 00m00s [148/307] xnnpack-0:0.0^git20240229.fcb 100% | 61.5 MiB/s | 693.2 KiB | 00m00s [149/307] cpuinfo-0:23.11.04-0.gitd6860 100% | 6.9 MiB/s | 42.5 KiB | 00m00s [150/307] pthreadpool-0:0.0^git20230829 100% | 5.6 MiB/s | 46.1 KiB | 00m00s [151/307] cmake-filesystem-0:3.30.5-1.f 100% | 3.4 MiB/s | 17.4 KiB | 00m00s [152/307] roctracer-0:6.2.0-2.fc41.x86_ 100% | 28.2 MiB/s | 288.8 KiB | 00m00s [153/307] sleef-0:3.7.0-1.fc41.x86_64 100% | 53.6 MiB/s | 823.4 KiB | 00m00s [154/307] valgrind-1:3.24.0-2.fc41.x86_ 100% | 156.4 MiB/s | 5.3 MiB | 00m00s [155/307] rocm-runtime-0:6.2.1-2.fc41.x 100% | 35.1 MiB/s | 539.6 KiB | 00m00s [156/307] libatomic-0:14.2.1-3.fc41.x86 100% | 6.9 MiB/s | 42.5 KiB | 00m00s [157/307] rocm-comgr-0:18-10.rocm6.2.1. 100% | 118.0 MiB/s | 2.8 MiB | 00m00s [158/307] rocrand-gfx1103-0:6.2.0-1.fc4 100% | 78.1 MiB/s | 19.2 MiB | 00m00s [159/307] rocrand-gfx1100-0:6.2.0-1.fc4 100% | 63.1 MiB/s | 19.3 MiB | 00m00s [160/307] rocrand-0:6.2.0-1.fc41.x86_64 100% | 151.6 MiB/s | 85.9 MiB | 00m01s [161/307] rocrand-gfx90a-0:6.2.0-1.fc41 100% | 56.3 MiB/s | 19.1 MiB | 00m00s [162/307] hsakmt-devel-0:1.0.6-46.rocm6 100% | 3.2 MiB/s | 36.6 KiB | 00m00s [163/307] hsakmt-0:1.0.6-46.rocm6.2.1.f 100% | 8.9 MiB/s | 73.1 KiB | 00m00s [164/307] rocrand-gfx942-0:6.2.0-1.fc41 100% | 57.9 MiB/s | 18.9 MiB | 00m00s [165/307] libdrm-0:2.4.124-1.fc41.x86_6 100% | 6.3 MiB/s | 148.5 KiB | 00m00s [166/307] perl-File-Copy-0:2.41-512.fc4 100% | 3.3 MiB/s | 20.1 KiB | 00m00s [167/307] environment-modules-0:5.4.0-2 100% | 47.0 MiB/s | 721.9 KiB | 00m00s [168/307] perl-File-Which-0:1.27-12.fc4 100% | 2.1 MiB/s | 21.7 KiB | 00m00s [169/307] hipcc-0:18-10.rocm6.2.1.fc41. 100% | 14.2 MiB/s | 131.1 KiB | 00m00s [170/307] numactl-libs-0:2.0.19-1.fc41. 100% | 3.4 MiB/s | 31.0 KiB | 00m00s [171/307] rocm-hip-0:6.2.1-2.fc41.x86_6 100% | 81.6 MiB/s | 9.4 MiB | 00m00s [172/307] rocm-core-0:6.2.0-1.fc41.x86_ 100% | 721.2 KiB/s | 13.0 KiB | 00m00s [173/307] clang18-devel-0:18.1.8-5.fc41 100% | 98.3 MiB/s | 3.0 MiB | 00m00s [174/307] llvm18-devel-0:18.1.8-4.fc41. 100% | 127.0 MiB/s | 3.7 MiB | 00m00s [175/307] lld18-libs-0:18.1.8-6.fc41.x8 100% | 64.3 MiB/s | 1.5 MiB | 00m00s [176/307] rocfft-0:6.2.1-2.fc41.x86_64 100% | 92.2 MiB/s | 3.0 MiB | 00m00s [177/307] llvm18-libs-0:18.1.8-4.fc41.x 100% | 192.9 MiB/s | 28.0 MiB | 00m00s [178/307] clang18-libs-0:18.1.8-5.fc41. 100% | 109.9 MiB/s | 21.8 MiB | 00m00s [179/307] rccl-data-0:6.2.1-1.fc41.noar 100% | 2.5 MiB/s | 134.6 KiB | 00m00s [180/307] rocm-smi-0:6.2.1-1.fc41.x86_6 100% | 13.6 MiB/s | 557.3 KiB | 00m00s [181/307] python3-mpmath-0:1.3.0-10.fc4 100% | 80.8 MiB/s | 1.1 MiB | 00m00s [182/307] pybind11-devel-0:2.13.6-1.fc4 100% | 21.6 MiB/s | 176.8 KiB | 00m00s [183/307] flexiblas-netlib-0:3.4.4-3.fc 100% | 128.7 MiB/s | 3.2 MiB | 00m00s [184/307] protobuf-0:3.19.6-10.fc41.x86 100% | 66.6 MiB/s | 1.0 MiB | 00m00s [185/307] protobuf-compiler-0:3.19.6-10 100% | 68.8 MiB/s | 775.4 KiB | 00m00s [186/307] zlib-ng-compat-devel-0:2.1.7- 100% | 7.4 MiB/s | 38.1 KiB | 00m00s [187/307] openmpi-0:5.0.5-2.fc41.x86_64 100% | 100.7 MiB/s | 2.0 MiB | 00m00s [188/307] pmix-0:4.2.8-3.fc41.x86_64 100% | 49.8 MiB/s | 662.9 KiB | 00m00s [189/307] rpm-mpi-hooks-0:8-10.fc41.noa 100% | 1.3 MiB/s | 10.7 KiB | 00m00s [190/307] prrte-0:3.0.6-1.fc41.x86_64 100% | 9.2 MiB/s | 56.8 KiB | 00m00s [191/307] ucx-0:1.17.0-3.fc41.x86_64 100% | 62.3 MiB/s | 829.8 KiB | 00m00s [192/307] hwloc-libs-0:2.11.2-1.fc41.x8 100% | 109.9 MiB/s | 2.1 MiB | 00m00s [193/307] libfabric-0:1.22.0-1.fc41.x86 100% | 78.0 MiB/s | 1.4 MiB | 00m00s [194/307] libgfortran-0:14.2.1-3.fc41.x 100% | 76.4 MiB/s | 939.1 KiB | 00m00s [195/307] libpsm2-0:11.2.230-5.fc41.x86 100% | 21.7 MiB/s | 200.0 KiB | 00m00s [196/307] libquadmath-0:14.2.1-3.fc41.x 100% | 33.0 MiB/s | 202.5 KiB | 00m00s [197/307] orangefs-0:2.9.8-12.fc41.x86_ 100% | 91.7 MiB/s | 1.8 MiB | 00m00s [198/307] onnx-libs-0:1.15.0-4.fc41.x86 100% | 71.7 MiB/s | 880.9 KiB | 00m00s [199/307] vim-filesystem-2:9.1.984-1.fc 100% | 3.2 MiB/s | 16.4 KiB | 00m00s [200/307] emacs-filesystem-1:30.0-3.fc4 100% | 1.7 MiB/s | 7.1 KiB | 00m00s [201/307] miopen-0:6.2.1-1.fc41.x86_64 100% | 171.2 MiB/s | 25.5 MiB | 00m00s [202/307] miopen-gfx1100-0:6.2.1-1.fc41 100% | 39.2 MiB/s | 6.4 MiB | 00m00s [203/307] miopen-gfx1103-0:6.2.1-1.fc41 100% | 31.3 MiB/s | 6.4 MiB | 00m00s [204/307] rccl-0:6.2.1-1.fc41.x86_64 100% | 120.8 MiB/s | 121.9 MiB | 00m01s [205/307] miopen-gfx90a-0:6.2.1-1.fc41. 100% | 37.3 MiB/s | 6.4 MiB | 00m00s [206/307] boost-filesystem-0:1.83.0-8.f 100% | 8.4 MiB/s | 68.6 KiB | 00m00s [207/307] miopen-gfx942-0:6.2.1-1.fc41. 100% | 35.1 MiB/s | 6.4 MiB | 00m00s [208/307] lapack-0:3.12.0-7.fc41.x86_64 100% | 189.3 MiB/s | 10.8 MiB | 00m00s [209/307] libedit-devel-0:3.1-54.202501 100% | 4.0 MiB/s | 40.7 KiB | 00m00s [210/307] llvm18-0:18.1.8-4.fc41.x86_64 100% | 142.9 MiB/s | 26.9 MiB | 00m00s [211/307] llvm18-googletest-0:18.1.8-4. 100% | 28.8 MiB/s | 383.1 KiB | 00m00s [212/307] llvm18-static-0:18.1.8-4.fc41 100% | 222.1 MiB/s | 38.0 MiB | 00m00s [213/307] llvm18-test-0:18.1.8-4.fc41.x 100% | 53.0 MiB/s | 651.0 KiB | 00m00s [214/307] clang-resource-filesystem-0:1 100% | 1.3 MiB/s | 19.3 KiB | 00m00s [215/307] libomp-0:19.1.5-1.fc41.x86_64 100% | 173.7 MiB/s | 13.6 MiB | 00m00s [216/307] llvm-libs-0:19.1.5-1.fc41.x86 100% | 214.2 MiB/s | 30.0 MiB | 00m00s [217/307] libibverbs-0:51.0-5.fc41.x86_ 100% | 34.9 MiB/s | 429.4 KiB | 00m00s [218/307] libnl3-0:3.11.0-1.fc41.x86_64 100% | 17.2 MiB/s | 352.9 KiB | 00m00s [219/307] librdmacm-0:51.0-5.fc41.x86_6 100% | 6.4 MiB/s | 72.5 KiB | 00m00s [220/307] ncurses-devel-0:6.5-2.2024062 100% | 35.4 MiB/s | 543.2 KiB | 00m00s [221/307] libpciaccess-0:0.16-13.fc41.x 100% | 3.2 MiB/s | 26.5 KiB | 00m00s [222/307] blas-0:3.12.0-7.fc41.x86_64 100% | 39.9 MiB/s | 694.6 KiB | 00m00s [223/307] hipsparse-0:6.2.0-1.fc41.x86_ 100% | 6.9 MiB/s | 71.0 KiB | 00m00s [224/307] hipsparse-gfx1100-0:6.2.0-1.f 100% | 2.1 MiB/s | 59.7 KiB | 00m00s [225/307] hipsparse-gfx1103-0:6.2.0-1.f 100% | 1.5 MiB/s | 59.8 KiB | 00m00s [226/307] hipsparse-gfx90a-0:6.2.0-1.fc 100% | 2.2 MiB/s | 59.7 KiB | 00m00s [227/307] hipsparse-gfx942-0:6.2.0-1.fc 100% | 1.6 MiB/s | 59.7 KiB | 00m00s [228/307] hipsolver-0:6.2.0-1.fc41.x86_ 100% | 9.4 MiB/s | 86.5 KiB | 00m00s [229/307] hipsolver-gfx1100-0:6.2.0-1.f 100% | 2.8 MiB/s | 73.9 KiB | 00m00s [230/307] hipsolver-gfx1103-0:6.2.0-1.f 100% | 1.9 MiB/s | 73.9 KiB | 00m00s [231/307] hipsolver-gfx90a-0:6.2.0-1.fc 100% | 2.2 MiB/s | 73.9 KiB | 00m00s [232/307] magma-0:2.8.0-2.fc41.x86_64 100% | 193.3 MiB/s | 275.4 MiB | 00m01s [233/307] hipsolver-gfx942-0:6.2.0-1.fc 100% | 267.7 KiB/s | 73.9 KiB | 00m00s [234/307] hiprand-0:6.2.0-1.fc41.x86_64 100% | 2.9 MiB/s | 20.5 KiB | 00m00s [235/307] hiprand-gfx1100-0:6.2.0-1.fc4 100% | 732.4 KiB/s | 13.2 KiB | 00m00s [236/307] hiprand-gfx1103-0:6.2.0-1.fc4 100% | 944.1 KiB/s | 13.2 KiB | 00m00s [237/307] hiprand-gfx90a-0:6.2.0-1.fc41 100% | 941.4 KiB/s | 13.2 KiB | 00m00s [238/307] suitesparse-0:7.7.0-2.fc41.x8 100% | 225.1 MiB/s | 19.1 MiB | 00m00s [239/307] hiprand-gfx942-0:6.2.0-1.fc41 100% | 487.8 KiB/s | 13.2 KiB | 00m00s [240/307] perl-lib-0:0.65-512.fc41.x86_ 100% | 2.9 MiB/s | 14.9 KiB | 00m00s [241/307] hipfft-0:6.2.1-1.fc41.x86_64 100% | 6.4 MiB/s | 52.2 KiB | 00m00s [242/307] rocminfo-0:6.2.1-1.fc41.x86_6 100% | 5.5 MiB/s | 39.2 KiB | 00m00s [243/307] rocm-device-libs-0:18-10.rocm 100% | 49.5 MiB/s | 557.8 KiB | 00m00s [244/307] compiler-rt18-0:18.1.8-3.fc41 100% | 105.4 MiB/s | 2.3 MiB | 00m00s [245/307] hipblas-0:6.2.0-1.fc41.x86_64 100% | 13.3 MiB/s | 176.9 KiB | 00m00s [246/307] hipblas-gfx1100-0:6.2.0-1.fc4 100% | 7.2 MiB/s | 155.5 KiB | 00m00s [247/307] hipblas-gfx1103-0:6.2.0-1.fc4 100% | 4.1 MiB/s | 155.3 KiB | 00m00s [248/307] hipblaslt-0:6.2.0-3.fc41.x86_ 100% | 247.3 MiB/s | 31.4 MiB | 00m00s [249/307] hipblas-gfx90a-0:6.2.0-1.fc41 100% | 4.6 MiB/s | 155.4 KiB | 00m00s [250/307] hipblas-gfx942-0:6.2.0-1.fc41 100% | 6.6 MiB/s | 155.4 KiB | 00m00s [251/307] libmpc-0:1.3.1-6.fc41.x86_64 100% | 7.7 MiB/s | 71.1 KiB | 00m00s [252/307] libquadmath-devel-0:14.2.1-3. 100% | 4.8 MiB/s | 44.3 KiB | 00m00s [253/307] libstdc++-devel-0:14.2.1-3.fc 100% | 113.8 MiB/s | 2.6 MiB | 00m00s [254/307] glibc-devel-0:2.40-17.fc41.x8 100% | 26.8 MiB/s | 548.1 KiB | 00m00s [255/307] gcc-0:14.2.1-3.fc41.x86_64 100% | 236.6 MiB/s | 36.9 MiB | 00m00s [256/307] make-1:4.4.1-8.fc41.x86_64 100% | 51.4 MiB/s | 579.1 KiB | 00m00s [257/307] cpp-0:14.2.1-3.fc41.x86_64 100% | 123.1 MiB/s | 11.9 MiB | 00m00s [258/307] foxi-0:1.4.1^git20210526.c278 100% | 742.5 KiB/s | 11.9 KiB | 00m00s [259/307] rocblas-0:6.2.1-1.fc41.x86_64 100% | 139.8 MiB/s | 484.0 MiB | 00m03s [260/307] flexiblas-0:3.4.4-3.fc41.x86_ 100% | 57.3 KiB/s | 25.3 KiB | 00m00s [261/307] fmt-0:11.0.2-2.fc41.x86_64 100% | 227.3 KiB/s | 100.2 KiB | 00m00s [262/307] flexiblas-openblas-openmp-0:3 100% | 4.2 MiB/s | 17.2 KiB | 00m00s [263/307] man-db-0:2.12.1-2.fc41.x86_64 100% | 96.6 MiB/s | 1.3 MiB | 00m00s [264/307] procps-ng-0:4.0.4-4.fc41.x86_ 100% | 29.2 MiB/s | 359.0 KiB | 00m00s [265/307] flexiblas-netlib64-0:3.4.4-3. 100% | 147.3 MiB/s | 3.1 MiB | 00m00s [266/307] clang18-resource-filesystem-0 100% | 1.9 MiB/s | 13.5 KiB | 00m00s [267/307] tcl-1:8.6.14-2.fc41.x86_64 100% | 96.9 MiB/s | 1.1 MiB | 00m00s [268/307] libuv-1:1.49.2-1.fc41.x86_64 100% | 42.7 MiB/s | 262.1 KiB | 00m00s [269/307] jsoncpp-0:1.9.5-8.fc41.x86_64 100% | 16.2 MiB/s | 99.3 KiB | 00m00s [270/307] cmake-data-0:3.30.5-1.fc41.no 100% | 151.4 MiB/s | 1.8 MiB | 00m00s [271/307] rhash-0:1.4.4-2.fc41.x86_64 100% | 31.9 MiB/s | 196.0 KiB | 00m00s [272/307] clang18-0:18.1.8-5.fc41.x86_6 100% | 10.2 MiB/s | 72.9 KiB | 00m00s [273/307] boost-atomic-0:1.83.0-8.fc41. 100% | 3.7 MiB/s | 18.9 KiB | 00m00s [274/307] boost-system-0:1.83.0-8.fc41. 100% | 2.9 MiB/s | 14.7 KiB | 00m00s [275/307] kmod-0:33-1.fc41.x86_64 100% | 17.2 MiB/s | 123.0 KiB | 00m00s [276/307] python3-peachpy-0:0.2.0-2.fc4 100% | 56.0 MiB/s | 688.7 KiB | 00m00s [277/307] lld18-devel-0:18.1.8-6.fc41.x 100% | 4.0 MiB/s | 24.7 KiB | 00m00s [278/307] munge-libs-0:0.5.16-3.fc41.x8 100% | 3.5 MiB/s | 21.4 KiB | 00m00s [279/307] prrte-libs-0:3.0.6-1.fc41.x86 100% | 37.3 MiB/s | 534.1 KiB | 00m00s [280/307] tcsh-0:6.24.13-1.fc41.x86_64 100% | 44.9 MiB/s | 459.5 KiB | 00m00s [281/307] ncurses-c++-libs-0:6.5-2.2024 100% | 5.3 MiB/s | 37.8 KiB | 00m00s [282/307] lld18-0:18.1.8-6.fc41.x86_64 100% | 4.4 MiB/s | 26.9 KiB | 00m00s [283/307] libpipeline-0:1.5.7-6.fc41.x8 100% | 6.3 MiB/s | 51.9 KiB | 00m00s [284/307] kernel-headers-0:6.12.4-200.f 100% | 113.8 MiB/s | 1.5 MiB | 00m00s [285/307] hwdata-0:0.391-1.fc41.noarch 100% | 95.8 MiB/s | 1.6 MiB | 00m00s [286/307] clang18-tools-extra-0:18.1.8- 100% | 225.9 MiB/s | 19.7 MiB | 00m00s [287/307] libxcrypt-devel-0:4.4.37-4.fc 100% | 1.3 MiB/s | 28.3 KiB | 00m00s [288/307] flexiblas-openblas-openmp64-0 100% | 2.8 MiB/s | 17.2 KiB | 00m00s [289/307] torque-libs-0:6.1.3-13.fc41.x 100% | 20.4 MiB/s | 187.8 KiB | 00m00s [290/307] openblas-0:0.3.26-5.fc41.x86_ 100% | 7.6 MiB/s | 38.7 KiB | 00m00s [291/307] openblas-openmp-0:0.3.26-5.fc 100% | 117.8 MiB/s | 5.1 MiB | 00m00s [292/307] munge-0:0.5.16-3.fc41.x86_64 100% | 9.7 MiB/s | 129.1 KiB | 00m00s [293/307] logrotate-0:3.22.0-2.fc41.x86 100% | 9.3 MiB/s | 75.9 KiB | 00m00s [294/307] openblas-openmp64-0:0.3.26-5. 100% | 164.7 MiB/s | 4.9 MiB | 00m00s [295/307] systemd-pam-0:256.10-1.fc41.x 100% | 30.7 MiB/s | 376.9 KiB | 00m00s [296/307] dbus-1:1.14.10-4.fc41.x86_64 100% | 1.1 MiB/s | 7.9 KiB | 00m00s [297/307] libseccomp-0:2.5.5-2.fc41.x86 100% | 11.4 MiB/s | 70.2 KiB | 00m00s [298/307] dbus-broker-0:36-4.fc41.x86_6 100% | 27.9 MiB/s | 171.7 KiB | 00m00s [299/307] systemd-0:256.10-1.fc41.x86_6 100% | 156.8 MiB/s | 5.3 MiB | 00m00s [300/307] dbus-common-1:1.14.10-4.fc41. 100% | 1.6 MiB/s | 14.7 KiB | 00m00s [301/307] systemd-rpm-macros-0:256.10-1 100% | 6.5 MiB/s | 33.3 KiB | 00m00s [302/307] annobin-plugin-gcc-0:12.69-1. 100% | 94.8 MiB/s | 971.0 KiB | 00m00s [303/307] gcc-plugin-annobin-0:14.2.1-3 100% | 6.7 MiB/s | 55.1 KiB | 00m00s [304/307] cmake-rpm-macros-0:3.30.5-1.f 100% | 2.7 MiB/s | 16.8 KiB | 00m00s [305/307] annobin-docs-0:12.69-1.fc41.n 100% | 11.2 MiB/s | 91.8 KiB | 00m00s [306/307] rocsolver-0:6.2.0-1.fc41.x86_ 100% | 290.9 MiB/s | 115.5 MiB | 00m00s [307/307] rocsparse-0:6.2.1-1.fc41.x86_ 100% | 265.6 MiB/s | 450.3 MiB | 00m02s -------------------------------------------------------------------------------- [307/307] Total 100% | 317.1 MiB/s | 2.1 GiB | 00m07s Running transaction [ 1/309] Verify package files 100% | 48.0 B/s | 307.0 B | 00m06s [ 2/309] Prepare transaction 100% | 1.3 KiB/s | 307.0 B | 00m00s [ 3/309] Installing cmake-filesystem-0 100% | 7.1 MiB/s | 7.3 KiB | 00m00s [ 4/309] Installing libgfortran-0:14.2 100% | 380.5 MiB/s | 3.0 MiB | 00m00s [ 5/309] Installing libquadmath-0:14.2 100% | 319.4 MiB/s | 327.1 KiB | 00m00s [ 6/309] Installing hwloc-libs-0:2.11. 100% | 477.0 MiB/s | 2.9 MiB | 00m00s [ 7/309] Installing numactl-libs-0:2.0 100% | 0.0 B/s | 61.8 KiB | 00m00s [ 8/309] Installing libmpc-0:1.3.1-6.f 100% | 162.3 MiB/s | 166.2 KiB | 00m00s [ 9/309] Installing libedit-0:3.1-54.2 100% | 235.3 MiB/s | 241.0 KiB | 00m00s [ 10/309] Installing llvm18-libs-0:18.1 100% | 388.7 MiB/s | 113.5 MiB | 00m00s [ 11/309] Installing lld18-libs-0:18.1. 100% | 350.6 MiB/s | 5.3 MiB | 00m00s [ 12/309] Installing python-rpm-macros- 100% | 0.0 B/s | 22.8 KiB | 00m00s [ 13/309] Installing munge-libs-0:0.5.1 100% | 0.0 B/s | 32.9 KiB | 00m00s [ 14/309] Installing pmix-0:4.2.8-3.fc4 100% | 336.0 MiB/s | 2.0 MiB | 00m00s [ 15/309] Installing expat-0:2.6.4-1.fc 100% | 288.1 MiB/s | 295.0 KiB | 00m00s [ 16/309] Installing clang18-resource-f 100% | 980.5 KiB/s | 1.0 KiB | 00m00s [ 17/309] Installing clang18-libs-0:18. 100% | 442.1 MiB/s | 102.1 MiB | 00m00s [ 18/309] Installing rocm-comgr-0:18-10 100% | 405.4 MiB/s | 8.9 MiB | 00m00s [ 19/309] Installing libnl3-0:3.11.0-1. 100% | 261.9 MiB/s | 1.0 MiB | 00m00s [ 20/309] Installing libibverbs-0:51.0- 100% | 241.9 MiB/s | 1.2 MiB | 00m00s [ 21/309] Installing emacs-filesystem-1 100% | 0.0 B/s | 544.0 B | 00m00s [ 22/309] Installing vim-filesystem-2:9 100% | 4.6 MiB/s | 4.7 KiB | 00m00s [ 23/309] Installing protobuf-0:3.19.6- 100% | 361.9 MiB/s | 3.3 MiB | 00m00s [ 24/309] Installing python3-rpm-macros 100% | 0.0 B/s | 6.7 KiB | 00m00s [ 25/309] Installing libpsm2-0:11.2.230 100% | 216.6 MiB/s | 443.6 KiB | 00m00s [ 26/309] Installing blas-0:3.12.0-7.fc 100% | 201.6 MiB/s | 1.8 MiB | 00m00s [ 27/309] Installing openblas-0:0.3.26- 100% | 11.9 MiB/s | 97.8 KiB | 00m00s [ 28/309] Installing groff-base-0:1.23. 100% | 176.1 MiB/s | 3.9 MiB | 00m00s [ 29/309] Installing jsoncpp-0:1.9.5-8. 100% | 124.5 MiB/s | 254.9 KiB | 00m00s [ 30/309] Installing less-0:661-2.fc41. 100% | 199.5 MiB/s | 408.6 KiB | 00m00s [ 31/309] Installing make-1:4.4.1-8.fc4 100% | 225.0 MiB/s | 1.8 MiB | 00m00s [ 32/309] Installing libstdc++-devel-0: 100% | 353.6 MiB/s | 15.6 MiB | 00m00s [ 33/309] Installing pthreadpool-0:0.0^ 100% | 111.6 MiB/s | 114.3 KiB | 00m00s [ 34/309] Installing cpuinfo-0:23.11.04 100% | 114.0 MiB/s | 116.7 KiB | 00m00s [ 35/309] Installing xnnpack-0:0.0^git2 100% | 313.9 MiB/s | 1.9 MiB | 00m00s [ 36/309] Installing openblas-openmp-0: 100% | 596.5 MiB/s | 39.4 MiB | 00m00s [ 37/309] Installing flexiblas-0:3.4.4- 100% | 0.0 B/s | 49.7 KiB | 00m00s [ 38/309] Installing flexiblas-openblas 100% | 0.0 B/s | 44.1 KiB | 00m00s [ 39/309] Installing flexiblas-netlib-0 100% | 395.0 MiB/s | 10.7 MiB | 00m00s [ 40/309] Installing suitesparse-0:7.7. 100% | 524.2 MiB/s | 140.0 MiB | 00m00s [ 41/309] Installing openblas-openmp64- 100% | 581.3 MiB/s | 39.5 MiB | 00m00s [ 42/309] Installing flexiblas-netlib64 100% | 384.8 MiB/s | 10.8 MiB | 00m00s [ 43/309] Installing flexiblas-openblas 100% | 4.3 MiB/s | 44.2 KiB | 00m00s [ 44/309] Installing lapack-0:3.12.0-7. 100% | 102.6 MiB/s | 17.9 MiB | 00m00s [ 45/309] Installing pyproject-rpm-macr 100% | 113.0 MiB/s | 115.7 KiB | 00m00s [ 46/309] Installing protobuf-compiler- 100% | 352.3 MiB/s | 2.5 MiB | 00m00s [ 47/309] Installing onnx-libs-0:1.15.0 100% | 398.7 MiB/s | 3.2 MiB | 00m00s [ 48/309] Installing librdmacm-0:51.0-5 100% | 152.5 MiB/s | 156.2 KiB | 00m00s [ 49/309] Installing libfabric-0:1.22.0 100% | 397.2 MiB/s | 5.2 MiB | 00m00s [ 50/309] Installing lld18-0:18.1.8-6.f 100% | 134.1 MiB/s | 137.4 KiB | 00m00s [ 51/309] Installing lld18-devel-0:18.1 100% | 42.0 MiB/s | 43.0 KiB | 00m00s [ 52/309] Installing llvm-libs-0:19.1.5 100% | 386.7 MiB/s | 123.0 MiB | 00m00s [ 53/309] Installing libomp-0:19.1.5-1. 100% | 371.5 MiB/s | 49.8 MiB | 00m00s [ 54/309] Installing cpp-0:14.2.1-3.fc4 100% | 364.6 MiB/s | 35.0 MiB | 00m00s [ 55/309] Installing rocprim-devel-0:6. 100% | 455.7 MiB/s | 3.6 MiB | 00m00s [ 56/309] Installing zlib-ng-compat-dev 100% | 105.8 MiB/s | 108.3 KiB | 00m00s [ 57/309] Installing annobin-docs-0:12. 100% | 0.0 B/s | 98.8 KiB | 00m00s [ 58/309] Installing dbus-common-1:1.14 100% | 713.2 KiB/s | 13.6 KiB | 00m00s [ 59/309] Installing dbus-broker-0:36-4 100% | 47.0 MiB/s | 385.3 KiB | 00m00s [ 60/309] Installing dbus-1:1.14.10-4.f 100% | 0.0 B/s | 124.0 B | 00m00s [ 61/309] Installing libseccomp-0:2.5.5 100% | 171.1 MiB/s | 175.2 KiB | 00m00s [ 62/309] Installing systemd-pam-0:256. 100% | 212.7 MiB/s | 1.1 MiB | 00m00s [ 63/309] Installing systemd-0:256.10-1 100% | 112.4 MiB/s | 16.9 MiB | 00m00s >>> Running post-install scriptlet: systemd-0:256.10-1.fc41.x86_64 >>> Finished post-install scriptlet: systemd-0:256.10-1.fc41.x86_64 >>> Scriptlet output: >>> Creating group 'systemd-journal' with GID 190. >>> Creating group 'systemd-oom' with GID 999. >>> Creating user 'systemd-oom' (systemd Userspace OOM Killer) with UID 999 and >>> [ 64/309] Installing logrotate-0:3.22.0 100% | 6.1 MiB/s | 155.7 KiB | 00m00s >>> Running post-install scriptlet: logrotate-0:3.22.0-2.fc41.x86_64 >>> Finished post-install scriptlet: logrotate-0:3.22.0-2.fc41.x86_64 >>> Scriptlet output: >>> Created symlink '/etc/systemd/system/timers.target.wants/logrotate.timer' ↠>>> [ 65/309] Installing munge-0:0.5.16-3.f 100% | 24.6 MiB/s | 352.4 KiB | 00m00s [ 66/309] Installing torque-libs-0:6.1. 100% | 148.1 MiB/s | 455.0 KiB | 00m00s [ 67/309] Installing prrte-libs-0:3.0.6 100% | 241.1 MiB/s | 1.7 MiB | 00m00s [ 68/309] Installing prrte-0:3.0.6-1.fc 100% | 174.7 MiB/s | 178.9 KiB | 00m00s [ 69/309] Installing ncurses-0:6.5-2.20 100% | 206.3 MiB/s | 633.9 KiB | 00m00s [ 70/309] Installing perl-Digest-0:1.20 100% | 0.0 B/s | 37.1 KiB | 00m00s [ 71/309] Installing perl-B-0:1.89-512. 100% | 244.8 MiB/s | 501.4 KiB | 00m00s [ 72/309] Installing perl-FileHandle-0: 100% | 0.0 B/s | 9.8 KiB | 00m00s [ 73/309] Installing perl-Digest-MD5-0: 100% | 60.2 MiB/s | 61.7 KiB | 00m00s [ 74/309] Installing perl-MIME-Base32-0 100% | 0.0 B/s | 32.2 KiB | 00m00s [ 75/309] Installing perl-Data-Dumper-0 100% | 110.9 MiB/s | 113.6 KiB | 00m00s [ 76/309] Installing perl-libnet-0:3.15 100% | 143.9 MiB/s | 294.7 KiB | 00m00s [ 77/309] Installing perl-IO-Socket-IP- 100% | 99.8 MiB/s | 102.2 KiB | 00m00s [ 78/309] Installing perl-AutoLoader-0: 100% | 0.0 B/s | 20.9 KiB | 00m00s [ 79/309] Installing perl-URI-0:5.30-1. 100% | 87.7 MiB/s | 269.5 KiB | 00m00s [ 80/309] Installing perl-if-0:0.61.000 100% | 0.0 B/s | 6.2 KiB | 00m00s [ 81/309] Installing perl-Time-Local-2: 100% | 68.9 MiB/s | 70.6 KiB | 00m00s [ 82/309] Installing perl-locale-0:1.12 100% | 0.0 B/s | 6.9 KiB | 00m00s [ 83/309] Installing perl-File-Path-0:2 100% | 0.0 B/s | 64.5 KiB | 00m00s [ 84/309] Installing perl-Pod-Escapes-1 100% | 0.0 B/s | 25.9 KiB | 00m00s [ 85/309] Installing perl-Text-Tabs+Wra 100% | 23.3 MiB/s | 23.9 KiB | 00m00s [ 86/309] Installing perl-Net-SSLeay-0: 100% | 272.5 MiB/s | 1.4 MiB | 00m00s [ 87/309] Installing perl-IO-Socket-SSL 100% | 345.4 MiB/s | 707.4 KiB | 00m00s [ 88/309] Installing perl-POSIX-0:2.20- 100% | 230.8 MiB/s | 236.4 KiB | 00m00s [ 89/309] Installing perl-Class-Struct- 100% | 0.0 B/s | 25.9 KiB | 00m00s [ 90/309] Installing perl-File-Temp-1:0 100% | 160.2 MiB/s | 164.1 KiB | 00m00s [ 91/309] Installing perl-IPC-Open3-0:1 100% | 0.0 B/s | 23.3 KiB | 00m00s [ 92/309] Installing perl-Term-ANSIColo 100% | 96.9 MiB/s | 99.2 KiB | 00m00s [ 93/309] Installing perl-Term-Cap-0:1. 100% | 0.0 B/s | 30.6 KiB | 00m00s [ 94/309] Installing perl-HTTP-Tiny-0:0 100% | 152.8 MiB/s | 156.4 KiB | 00m00s [ 95/309] Installing perl-Pod-Simple-1: 100% | 278.5 MiB/s | 570.5 KiB | 00m00s [ 96/309] Installing perl-Symbol-0:1.09 100% | 0.0 B/s | 7.2 KiB | 00m00s [ 97/309] Installing perl-Socket-4:2.03 100% | 123.1 MiB/s | 126.1 KiB | 00m00s [ 98/309] Installing perl-SelectSaver-0 100% | 0.0 B/s | 2.6 KiB | 00m00s [ 99/309] Installing perl-File-stat-0:1 100% | 0.0 B/s | 13.1 KiB | 00m00s [100/309] Installing perl-Pod-Perldoc-0 100% | 165.3 MiB/s | 169.3 KiB | 00m00s [101/309] Installing perl-podlators-1:6 100% | 313.9 MiB/s | 321.4 KiB | 00m00s [102/309] Installing perl-base-0:2.27-5 100% | 0.0 B/s | 12.9 KiB | 00m00s [103/309] Installing perl-overloading-0 100% | 0.0 B/s | 5.5 KiB | 00m00s [104/309] Installing perl-mro-0:1.29-51 100% | 0.0 B/s | 46.7 KiB | 00m00s [105/309] Installing perl-Fcntl-0:1.18- 100% | 0.0 B/s | 50.1 KiB | 00m00s [106/309] Installing perl-Text-ParseWor 100% | 0.0 B/s | 14.6 KiB | 00m00s [107/309] Installing perl-IO-0:1.55-512 100% | 151.7 MiB/s | 155.3 KiB | 00m00s [108/309] Installing perl-Pod-Usage-4:2 100% | 0.0 B/s | 86.3 KiB | 00m00s [109/309] Installing perl-Getopt-Std-0: 100% | 0.0 B/s | 11.7 KiB | 00m00s [110/309] Installing perl-Scalar-List-U 100% | 149.0 MiB/s | 152.6 KiB | 00m00s [111/309] Installing perl-MIME-Base64-0 100% | 47.2 MiB/s | 48.4 KiB | 00m00s [112/309] Installing perl-constant-0:1. 100% | 0.0 B/s | 27.4 KiB | 00m00s [113/309] Installing perl-parent-1:0.24 100% | 0.0 B/s | 10.7 KiB | 00m00s [114/309] Installing perl-Errno-0:1.38- 100% | 0.0 B/s | 8.8 KiB | 00m00s [115/309] Installing perl-vars-0:1.05-5 100% | 0.0 B/s | 4.3 KiB | 00m00s [116/309] Installing perl-overload-0:1. 100% | 0.0 B/s | 71.9 KiB | 00m00s [117/309] Installing perl-Storable-1:3. 100% | 228.5 MiB/s | 234.0 KiB | 00m00s [118/309] Installing perl-Getopt-Long-1 100% | 143.8 MiB/s | 147.2 KiB | 00m00s [119/309] Installing perl-File-Basename 100% | 0.0 B/s | 14.6 KiB | 00m00s [120/309] Installing perl-Carp-0:1.54-5 100% | 0.0 B/s | 47.7 KiB | 00m00s [121/309] Installing perl-Exporter-0:5. 100% | 0.0 B/s | 55.6 KiB | 00m00s [122/309] Installing perl-PathTools-0:3 100% | 180.2 MiB/s | 184.6 KiB | 00m00s [123/309] Installing perl-DynaLoader-0: 100% | 0.0 B/s | 32.5 KiB | 00m00s [124/309] Installing perl-Encode-4:3.21 100% | 362.9 MiB/s | 4.7 MiB | 00m00s [125/309] Installing perl-libs-4:5.40.0 100% | 276.8 MiB/s | 10.0 MiB | 00m00s [126/309] Installing perl-interpreter-4 100% | 121.1 MiB/s | 124.0 KiB | 00m00s [127/309] Installing perl-File-Copy-0:2 100% | 0.0 B/s | 20.2 KiB | 00m00s [128/309] Installing perl-File-Which-0: 100% | 0.0 B/s | 31.4 KiB | 00m00s [129/309] Installing perl-lib-0:0.65-51 100% | 8.7 MiB/s | 8.9 KiB | 00m00s [130/309] Installing kernel-headers-0:6 100% | 194.0 MiB/s | 6.6 MiB | 00m00s [131/309] Installing libxcrypt-devel-0: 100% | 16.0 MiB/s | 32.9 KiB | 00m00s [132/309] Installing glibc-devel-0:2.40 100% | 166.0 MiB/s | 2.3 MiB | 00m00s [133/309] Installing gcc-0:14.2.1-3.fc4 100% | 391.0 MiB/s | 104.4 MiB | 00m00s [134/309] Installing gcc-c++-0:14.2.1-3 100% | 353.4 MiB/s | 38.2 MiB | 00m00s [135/309] Installing clang18-0:18.1.8-5 100% | 316.0 MiB/s | 647.1 KiB | 00m00s [136/309] Installing libquadmath-devel- 100% | 0.0 B/s | 23.4 KiB | 00m00s [137/309] Installing gcc-gfortran-0:14. 100% | 358.3 MiB/s | 37.3 MiB | 00m00s [138/309] Installing libcbor-0:0.11.0-2 100% | 73.5 MiB/s | 75.3 KiB | 00m00s [139/309] Installing libfido2-0:1.15.0- 100% | 234.1 MiB/s | 239.7 KiB | 00m00s [140/309] Installing hwdata-0:0.391-1.f 100% | 493.0 MiB/s | 9.4 MiB | 00m00s [141/309] Installing libpciaccess-0:0.1 100% | 44.9 MiB/s | 46.0 KiB | 00m00s [142/309] Installing libdrm-0:2.4.124-1 100% | 196.1 MiB/s | 401.6 KiB | 00m00s [143/309] Installing hsakmt-0:1.0.6-46. 100% | 178.6 MiB/s | 182.9 KiB | 00m00s [144/309] Installing rocm-runtime-0:6.2 100% | 531.9 MiB/s | 2.7 MiB | 00m00s [145/309] Installing hsakmt-devel-0:1.0 100% | 109.9 MiB/s | 112.6 KiB | 00m00s [146/309] Installing rocm-runtime-devel 100% | 273.3 MiB/s | 559.6 KiB | 00m00s [147/309] Installing libpipeline-0:1.5. 100% | 10.2 MiB/s | 125.9 KiB | 00m00s [148/309] Installing man-db-0:2.12.1-2. 100% | 129.0 MiB/s | 2.8 MiB | 00m00s [149/309] Installing ncurses-c++-libs-0 100% | 53.0 MiB/s | 162.9 KiB | 00m00s [150/309] Installing ncurses-devel-0:6. 100% | 62.1 MiB/s | 1.0 MiB | 00m00s [151/309] Installing libedit-devel-0:3. 100% | 65.4 MiB/s | 67.0 KiB | 00m00s [152/309] Installing openssh-0:9.8p1-3. 100% | 356.5 MiB/s | 1.8 MiB | 00m00s [153/309] Installing openssh-clients-0: 100% | 104.8 MiB/s | 2.6 MiB | 00m00s [154/309] Installing tcsh-0:6.24.13-1.f 100% | 95.8 MiB/s | 1.2 MiB | 00m00s [155/309] Installing orangefs-0:2.9.8-1 100% | 282.8 MiB/s | 3.1 MiB | 00m00s [156/309] Installing python-pip-wheel-0 100% | 620.8 MiB/s | 1.2 MiB | 00m00s [157/309] Installing mpdecimal-0:2.5.1- 100% | 201.2 MiB/s | 206.0 KiB | 00m00s [158/309] Installing libb2-0:0.98.1-12. 100% | 21.1 MiB/s | 43.3 KiB | 00m00s [159/309] Installing tzdata-0:2024b-1.f 100% | 55.4 MiB/s | 1.9 MiB | 00m00s [160/309] Installing python3-libs-0:3.1 100% | 312.3 MiB/s | 40.6 MiB | 00m00s [161/309] Installing python3-0:3.13.1-2 100% | 31.6 MiB/s | 32.4 KiB | 00m00s [162/309] Installing python3-packaging- 100% | 211.5 MiB/s | 433.2 KiB | 00m00s [163/309] Installing python3-idna-0:3.7 100% | 196.0 MiB/s | 602.1 KiB | 00m00s [164/309] Installing llvm18-0:18.1.8-4. 100% | 380.5 MiB/s | 112.2 MiB | 00m00s [165/309] Installing cmake-rpm-macros-0 100% | 0.0 B/s | 8.1 KiB | 00m00s [166/309] Installing llvm18-test-0:18.1 100% | 324.3 MiB/s | 1.9 MiB | 00m00s [167/309] Installing python3-urllib3-0: 100% | 200.9 MiB/s | 1.0 MiB | 00m00s [168/309] Installing python3-rpm-genera 100% | 81.0 MiB/s | 82.9 KiB | 00m00s [169/309] Installing valgrind-1:3.24.0- 100% | 501.9 MiB/s | 30.1 MiB | 00m00s [170/309] Installing rocm-smi-0:6.2.1-1 100% | 393.5 MiB/s | 2.4 MiB | 00m00s [171/309] Installing python3-mpmath-0:1 100% | 306.6 MiB/s | 5.2 MiB | 00m00s [172/309] Installing python3-babel-0:2. 100% | 316.7 MiB/s | 28.5 MiB | 00m00s [173/309] Installing python3-docutils-0 100% | 245.5 MiB/s | 4.9 MiB | 00m00s [174/309] Installing python3-imagesize- 100% | 18.7 MiB/s | 38.3 KiB | 00m00s [175/309] Installing python3-pygments-0 100% | 264.5 MiB/s | 10.8 MiB | 00m00s [176/309] Installing python3-snowballst 100% | 290.7 MiB/s | 1.7 MiB | 00m00s [177/309] Installing python3-sphinx-the 100% | 45.3 MiB/s | 46.4 KiB | 00m00s [178/309] Installing python3-charset-no 100% | 158.5 MiB/s | 324.7 KiB | 00m00s [179/309] Installing python3-requests-0 100% | 162.1 MiB/s | 498.0 KiB | 00m00s [180/309] Installing compiler-rt18-0:18 100% | 597.1 MiB/s | 28.1 MiB | 00m00s [181/309] Installing clang18-tools-extr 100% | 437.7 MiB/s | 85.3 MiB | 00m00s [182/309] Installing clang18-devel-0:18 100% | 410.8 MiB/s | 23.8 MiB | 00m00s [183/309] Installing python3-markupsafe 100% | 60.1 MiB/s | 61.5 KiB | 00m00s [184/309] Installing python3-jinja2-0:3 100% | 364.7 MiB/s | 2.9 MiB | 00m00s [185/309] Installing python3-six-0:1.16 100% | 117.9 MiB/s | 120.7 KiB | 00m00s [186/309] Installing python3-peachpy-0: 100% | 604.8 MiB/s | 13.3 MiB | 00m00s [187/309] Installing kmod-0:33-1.fc41.x 100% | 117.2 MiB/s | 240.0 KiB | 00m00s [188/309] Installing rocminfo-0:6.2.1-1 100% | 81.1 MiB/s | 83.0 KiB | 00m00s [189/309] Installing boost-system-0:1.8 100% | 0.0 B/s | 17.3 KiB | 00m00s [190/309] Installing boost-atomic-0:1.8 100% | 0.0 B/s | 22.0 KiB | 00m00s [191/309] Installing boost-filesystem-0 100% | 145.1 MiB/s | 148.6 KiB | 00m00s [192/309] Installing rhash-0:1.4.4-2.fc 100% | 173.4 MiB/s | 355.1 KiB | 00m00s [193/309] Installing libuv-1:1.49.2-1.f 100% | 92.4 MiB/s | 567.8 KiB | 00m00s [194/309] Installing cmake-data-0:3.30. 100% | 123.9 MiB/s | 8.8 MiB | 00m00s [195/309] Installing cmake-0:3.30.5-1.f 100% | 338.3 MiB/s | 32.5 MiB | 00m00s [196/309] Installing pybind11-devel-0:2 100% | 278.6 MiB/s | 856.0 KiB | 00m00s [197/309] Installing tcl-1:8.6.14-2.fc4 100% | 301.9 MiB/s | 4.2 MiB | 00m00s [198/309] Installing procps-ng-0:4.0.4- 100% | 208.2 MiB/s | 1.0 MiB | 00m00s [199/309] Installing environment-module 100% | 100.9 MiB/s | 1.7 MiB | 00m00s [200/309] Installing rocm-rpm-macros-mo 100% | 15.4 MiB/s | 31.4 KiB | 00m00s [201/309] Installing rpm-mpi-hooks-0:8- 100% | 0.0 B/s | 9.7 KiB | 00m00s [202/309] Installing fmt-0:11.0.2-2.fc4 100% | 259.4 MiB/s | 265.6 KiB | 00m00s [203/309] Installing foxi-0:1.4.1^git20 100% | 0.0 B/s | 17.4 KiB | 00m00s [204/309] Installing clang-resource-fil 100% | 16.3 MiB/s | 16.7 KiB | 00m00s [205/309] Installing llvm18-static-0:18 100% | 490.3 MiB/s | 283.9 MiB | 00m01s [206/309] Installing llvm18-googletest- 100% | 275.9 MiB/s | 2.2 MiB | 00m00s [207/309] Installing llvm18-devel-0:18. 100% | 245.9 MiB/s | 24.6 MiB | 00m00s [208/309] Installing rocm-comgr-devel-0 100% | 101.8 MiB/s | 104.3 KiB | 00m00s [209/309] Installing rocm-device-libs-0 100% | 355.7 MiB/s | 3.2 MiB | 00m00s [210/309] Installing hipcc-0:18-10.rocm 100% | 321.3 MiB/s | 658.1 KiB | 00m00s [211/309] Installing rocm-hip-0:6.2.1-2 100% | 332.0 MiB/s | 22.9 MiB | 00m00s [212/309] Installing rocblas-0:6.2.1-1. 100% | 135.8 MiB/s | 10.0 GiB | 01m16s [213/309] Installing rocsparse-0:6.2.1- 100% | 108.2 MiB/s | 5.2 GiB | 00m49s [214/309] Installing rocsolver-0:6.2.0- 100% | 115.7 MiB/s | 2.3 GiB | 00m21s [215/309] Installing rocrand-0:6.2.0-1. 100% | 78.7 MiB/s | 422.0 MiB | 00m05s [216/309] Installing hipblas-0:6.2.0-1. 100% | 81.1 MiB/s | 3.9 MiB | 00m00s [217/309] Installing hipsparse-0:6.2.0- 100% | 24.0 MiB/s | 984.2 KiB | 00m00s [218/309] Installing rocfft-0:6.2.1-2.f 100% | 152.5 MiB/s | 54.3 MiB | 00m00s [219/309] Installing hipfft-0:6.2.1-1.f 100% | 16.3 MiB/s | 583.0 KiB | 00m00s [220/309] Installing magma-0:2.8.0-2.fc 100% | 95.3 MiB/s | 2.3 GiB | 00m25s [221/309] Installing hiprand-0:6.2.0-1. 100% | 6.5 MiB/s | 100.3 KiB | 00m00s [222/309] Installing hiprand-gfx1100-0: 100% | 3.4 MiB/s | 24.1 KiB | 00m00s [223/309] Installing hiprand-gfx1103-0: 100% | 3.4 MiB/s | 24.1 KiB | 00m00s [224/309] Installing hiprand-gfx90a-0:6 100% | 1.7 MiB/s | 24.1 KiB | 00m00s [225/309] Installing hiprand-gfx942-0:6 100% | 2.1 MiB/s | 24.1 KiB | 00m00s [226/309] Installing hipsolver-0:6.2.0- 100% | 115.9 MiB/s | 1.3 MiB | 00m00s [227/309] Installing hipsolver-gfx1100- 100% | 31.9 MiB/s | 326.2 KiB | 00m00s [228/309] Installing hipsolver-gfx1103- 100% | 26.5 MiB/s | 326.2 KiB | 00m00s [229/309] Installing hipsolver-gfx90a-0 100% | 35.4 MiB/s | 326.2 KiB | 00m00s [230/309] Installing hipsolver-gfx942-0 100% | 21.2 MiB/s | 326.2 KiB | 00m00s [231/309] Installing hipblas-gfx1100-0: 100% | 60.8 MiB/s | 996.1 KiB | 00m00s [232/309] Installing hipblas-gfx1103-0: 100% | 57.2 MiB/s | 996.0 KiB | 00m00s [233/309] Installing hipblas-gfx90a-0:6 100% | 88.4 MiB/s | 996.1 KiB | 00m00s [234/309] Installing hipblas-gfx942-0:6 100% | 60.8 MiB/s | 996.1 KiB | 00m00s [235/309] Installing hipsparse-gfx1100- 100% | 24.0 MiB/s | 245.8 KiB | 00m00s [236/309] Installing hipsparse-gfx1103- 100% | 26.7 MiB/s | 245.9 KiB | 00m00s [237/309] Installing hipsparse-gfx90a-0 100% | 26.7 MiB/s | 245.9 KiB | 00m00s [238/309] Installing hipsparse-gfx942-0 100% | 24.0 MiB/s | 245.9 KiB | 00m00s [239/309] Installing rocrand-gfx1100-0: 100% | 41.8 MiB/s | 44.8 MiB | 00m01s [240/309] Installing rocrand-gfx1103-0: 100% | 64.8 MiB/s | 44.8 MiB | 00m01s [241/309] Installing rocrand-gfx90a-0:6 100% | 72.6 MiB/s | 57.3 MiB | 00m01s [242/309] Installing rocrand-gfx942-0:6 100% | 71.5 MiB/s | 43.2 MiB | 00m01s [243/309] Installing rocrand-devel-0:6. 100% | 25.3 MiB/s | 3.8 MiB | 00m00s [244/309] Installing hipblaslt-0:6.2.0- 100% | 119.3 MiB/s | 882.8 MiB | 00m07s [245/309] Installing ucx-0:1.17.0-3.fc4 100% | 38.0 MiB/s | 2.4 MiB | 00m00s [246/309] Installing openmpi-0:5.0.5-2. 100% | 63.6 MiB/s | 7.0 MiB | 00m00s [247/309] Installing libyaml-0:0.2.5-15 100% | 9.5 MiB/s | 135.8 KiB | 00m00s [248/309] Installing rccl-data-0:6.2.1- 100% | 528.3 MiB/s | 16.9 MiB | 00m00s [249/309] Installing rccl-0:6.2.1-1.fc4 100% | 102.0 MiB/s | 1.1 GiB | 00m11s [250/309] Installing rocm-core-0:6.2.0- 100% | 1.7 MiB/s | 17.6 KiB | 00m00s [251/309] Installing libatomic-0:14.2.1 100% | 3.7 MiB/s | 41.5 KiB | 00m00s [252/309] Installing roctracer-0:6.2.0- 100% | 29.9 MiB/s | 1.0 MiB | 00m00s [253/309] Installing miopen-0:6.2.1-1.f 100% | 132.4 MiB/s | 1.4 GiB | 00m11s [254/309] Installing miopen-gfx1100-0:6 100% | 134.9 MiB/s | 356.0 MiB | 00m03s [255/309] Installing miopen-gfx1103-0:6 100% | 111.6 MiB/s | 356.0 MiB | 00m03s [256/309] Installing miopen-gfx90a-0:6. 100% | 120.2 MiB/s | 356.0 MiB | 00m03s [257/309] Installing miopen-gfx942-0:6. 100% | 95.0 MiB/s | 356.0 MiB | 00m04s [258/309] Installing sleef-0:3.7.0-1.fc 100% | 58.6 MiB/s | 2.2 MiB | 00m00s [259/309] Installing sleef-devel-0:3.7. 100% | 6.4 MiB/s | 274.4 KiB | 00m00s [260/309] Installing miopen-devel-0:6.2 100% | 144.4 MiB/s | 548.3 MiB | 00m04s [261/309] Installing roctracer-devel-0: 100% | 32.0 MiB/s | 426.5 KiB | 00m00s [262/309] Installing rocm-core-devel-0: 100% | 472.7 KiB/s | 5.7 KiB | 00m00s [263/309] Installing rccl-devel-0:6.2.1 100% | 4.9 MiB/s | 136.0 KiB | 00m00s [264/309] Installing python3-pyyaml-0:6 100% | 8.6 MiB/s | 804.9 KiB | 00m00s [265/309] Installing openmpi-devel-0:5. 100% | 31.6 MiB/s | 3.3 MiB | 00m00s [266/309] Installing hipblaslt-devel-0: 100% | 18.4 MiB/s | 169.3 KiB | 00m00s [267/309] Installing hiprand-devel-0:6. 100% | 502.9 KiB/s | 298.7 KiB | 00m01s [268/309] Installing hipsparse-devel-0: 100% | 23.2 MiB/s | 809.1 KiB | 00m00s [269/309] Installing hipblas-devel-0:6. 100% | 99.2 MiB/s | 2.8 MiB | 00m00s [270/309] Installing hipsolver-devel-0: 100% | 29.7 MiB/s | 699.8 KiB | 00m00s [271/309] Installing magma-devel-0:2.8. 100% | 114.3 MiB/s | 1.6 MiB | 00m00s [272/309] Installing hipfft-devel-0:6.2 100% | 18.2 MiB/s | 148.8 KiB | 00m00s [273/309] Installing rocfft-devel-0:6.2 100% | 11.1 MiB/s | 124.6 KiB | 00m00s [274/309] Installing rocblas-devel-0:6. 100% | 157.8 MiB/s | 2.5 MiB | 00m00s [275/309] Installing rocm-hip-devel-0:6 100% | 112.4 MiB/s | 2.6 MiB | 00m00s [276/309] Installing libomp-devel-0:19. 100% | 611.1 MiB/s | 31.2 MiB | 00m00s [277/309] Installing foxi-devel-0:1.4.1 100% | 16.8 MiB/s | 120.4 KiB | 00m00s [278/309] Installing fmt-devel-0:11.0.2 100% | 40.3 MiB/s | 578.4 KiB | 00m00s [279/309] Installing rocm-rpm-macros-0: 100% | 2.4 MiB/s | 19.7 KiB | 00m00s [280/309] Installing python3-pybind11-0 100% | 55.7 MiB/s | 912.2 KiB | 00m00s [281/309] Installing rocm-cmake-0:6.2.0 100% | 18.7 MiB/s | 134.3 KiB | 00m00s [282/309] Installing FP16-devel-0:1.0^g 100% | 1.7 MiB/s | 35.7 KiB | 00m00s [283/309] Installing python3-sphinx-1:7 100% | 41.9 MiB/s | 11.0 MiB | 00m00s [284/309] Installing python3-sympy-0:1. 100% | 89.0 MiB/s | 79.2 MiB | 00m01s [285/309] Installing valgrind-devel-1:3 100% | 49.1 MiB/s | 503.2 KiB | 00m00s [286/309] Installing python3-devel-0:3. 100% | 62.6 MiB/s | 1.8 MiB | 00m00s [287/309] Installing ninja-build-0:1.12 100% | 26.6 MiB/s | 435.8 KiB | 00m00s [288/309] Installing python3-filelock-0 100% | 8.6 MiB/s | 97.1 KiB | 00m00s [289/309] Installing python3-fsspec-0:2 100% | 47.7 MiB/s | 1.8 MiB | 00m00s [290/309] Installing python3-networkx-0 100% | 61.5 MiB/s | 17.2 MiB | 00m00s [291/309] Installing python3-numpy-1:1. 100% | 93.5 MiB/s | 44.1 MiB | 00m00s [292/309] Installing python3-setuptools 100% | 77.2 MiB/s | 7.3 MiB | 00m00s [293/309] Installing python3-typing-ext 100% | 28.5 MiB/s | 466.9 KiB | 00m00s [294/309] Installing annobin-plugin-gcc 100% | 3.5 MiB/s | 986.7 KiB | 00m00s [295/309] Installing gcc-plugin-annobin 100% | 2.7 MiB/s | 62.6 KiB | 00m00s [296/309] Installing protobuf-devel-0:3 100% | 95.2 MiB/s | 2.8 MiB | 00m00s [297/309] Installing hipcub-devel-0:6.2 100% | 67.5 MiB/s | 1.1 MiB | 00m00s [298/309] Installing onnx-devel-0:1.15. 100% | 60.8 MiB/s | 1.2 MiB | 00m00s [299/309] Installing flexiblas-devel-0: 100% | 250.7 MiB/s | 4.8 MiB | 00m00s [300/309] Installing xnnpack-devel-0:0. 100% | 36.7 MiB/s | 225.8 KiB | 00m00s [301/309] Installing cpuinfo-devel-0:23 100% | 10.9 MiB/s | 66.9 KiB | 00m00s [302/309] Installing pthreadpool-devel- 100% | 16.2 MiB/s | 99.8 KiB | 00m00s [303/309] Installing numactl-devel-0:2. 100% | 2.5 MiB/s | 27.6 KiB | 00m00s [304/309] Installing eigen3-devel-0:3.4 100% | 121.0 MiB/s | 8.5 MiB | 00m00s [305/309] Installing rocthrust-devel-0: 100% | 92.3 MiB/s | 5.4 MiB | 00m00s [306/309] Installing systemd-rpm-macros 100% | 2.2 MiB/s | 11.2 KiB | 00m00s [307/309] Installing psimd-devel-0:2020 100% | 15.1 MiB/s | 46.4 KiB | 00m00s [308/309] Installing pocketfft-devel-0: 100% | 39.3 MiB/s | 120.8 KiB | 00m00s [309/309] Installing fxdiv-devel-0:1.0^ 100% | 11.8 KiB/s | 17.7 KiB | 00m02s Warning: skipped OpenPGP checks for 307 packages from repositories: copr_base, http_kojipkgs_fedoraproject_org_repos_f41_build_latest_basearch Complete! Finish: build setup for python-torch-2.4.0-10.fc41.src.rpm Start: rpmbuild python-torch-2.4.0-10.fc41.src.rpm Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1725321600 Executing(%mkbuilddir): /bin/sh -e /var/tmp/rpm-tmp.wksPJI + umask 022 + cd /builddir/build/BUILD/python-torch-2.4.0-build + test -d /builddir/build/BUILD/python-torch-2.4.0-build + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w /builddir/build/BUILD/python-torch-2.4.0-build + /usr/bin/rm -rf /builddir/build/BUILD/python-torch-2.4.0-build + /usr/bin/mkdir -p /builddir/build/BUILD/python-torch-2.4.0-build + /usr/bin/mkdir -p /builddir/build/BUILD/python-torch-2.4.0-build/SPECPARTS + RPM_EC=0 ++ jobs -p + exit 0 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.bjy5pe + umask 022 + cd /builddir/build/BUILD/python-torch-2.4.0-build + cd /builddir/build/BUILD/python-torch-2.4.0-build + rm -rf pytorch-v2.4.0 + /usr/lib/rpm/rpmuncompress -x /builddir/build/SOURCES/pytorch-v2.4.0.tar.gz + STATUS=0 + '[' 0 -ne 0 ']' + cd pytorch-v2.4.0 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + /usr/bin/patch -p1 -s --fuzz=0 --no-backup-if-mismatch -f + /usr/lib/rpm/rpmuncompress /builddir/build/SOURCES/0001-no-third_party-foxi.patch + /usr/lib/rpm/rpmuncompress /builddir/build/SOURCES/0001-Add-cmake-option-USE_SYSTEM_FBGEMM.patch + /usr/bin/patch -p1 -s --fuzz=0 --no-backup-if-mismatch -f + /usr/lib/rpm/rpmuncompress /builddir/build/SOURCES/0001-cuda-hip-signatures.patch + /usr/bin/patch -p1 -s --fuzz=0 --no-backup-if-mismatch -f + /usr/lib/rpm/rpmuncompress /builddir/build/SOURCES/0001-silence-an-assert.patch + /usr/bin/patch -p1 -s --fuzz=0 --no-backup-if-mismatch -f + /usr/lib/rpm/rpmuncompress /builddir/build/SOURCES/0001-disable-use-of-aotriton.patch + /usr/bin/patch -p1 -s --fuzz=0 --no-backup-if-mismatch -f + /usr/lib/rpm/rpmuncompress /builddir/build/SOURCES/0001-include-fmt-ranges.h-for-using-fmt-join.patch + /usr/bin/patch -p1 -s --fuzz=0 --no-backup-if-mismatch -f + rm -rf torch.egg-info + tar xf /builddir/build/SOURCES/v23.3.3.tar.gz + rm -rf third_party/flatbuffers/BUILD.bazel third_party/flatbuffers/CHANGELOG.md third_party/flatbuffers/CMake third_party/flatbuffers/CMakeLists.txt third_party/flatbuffers/CONTRIBUTING.md third_party/flatbuffers/FlatBuffers.podspec third_party/flatbuffers/Formatters.md third_party/flatbuffers/LICENSE third_party/flatbuffers/Package.swift third_party/flatbuffers/Package@swift-5.5.swift third_party/flatbuffers/SECURITY.md third_party/flatbuffers/WORKSPACE third_party/flatbuffers/android third_party/flatbuffers/bazel third_party/flatbuffers/benchmarks third_party/flatbuffers/build_defs.bzl third_party/flatbuffers/composer.json third_party/flatbuffers/conan third_party/flatbuffers/conanfile.py third_party/flatbuffers/dart third_party/flatbuffers/docs third_party/flatbuffers/examples third_party/flatbuffers/go third_party/flatbuffers/grpc third_party/flatbuffers/include third_party/flatbuffers/java third_party/flatbuffers/js third_party/flatbuffers/kotlin third_party/flatbuffers/lobster third_party/flatbuffers/lua third_party/flatbuffers/mjs third_party/flatbuffers/net third_party/flatbuffers/nim third_party/flatbuffers/package.json third_party/flatbuffers/php third_party/flatbuffers/python third_party/flatbuffers/readme.md third_party/flatbuffers/reflection third_party/flatbuffers/rust third_party/flatbuffers/samples third_party/flatbuffers/scripts third_party/flatbuffers/snap third_party/flatbuffers/src third_party/flatbuffers/swift third_party/flatbuffers/swift.swiftformat third_party/flatbuffers/tests third_party/flatbuffers/ts third_party/flatbuffers/tsconfig.json third_party/flatbuffers/tsconfig.mjs.json third_party/flatbuffers/typescript.bzl third_party/flatbuffers/yarn.lock + cp -r flatbuffers-23.3.3/BUILD.bazel flatbuffers-23.3.3/CHANGELOG.md flatbuffers-23.3.3/CMake flatbuffers-23.3.3/CMakeLists.txt flatbuffers-23.3.3/CONTRIBUTING.md flatbuffers-23.3.3/FlatBuffers.podspec flatbuffers-23.3.3/Formatters.md flatbuffers-23.3.3/LICENSE flatbuffers-23.3.3/Package.swift flatbuffers-23.3.3/Package@swift-5.5.swift flatbuffers-23.3.3/SECURITY.md flatbuffers-23.3.3/WORKSPACE flatbuffers-23.3.3/android flatbuffers-23.3.3/bazel flatbuffers-23.3.3/benchmarks flatbuffers-23.3.3/build_defs.bzl flatbuffers-23.3.3/composer.json flatbuffers-23.3.3/conan flatbuffers-23.3.3/conanfile.py flatbuffers-23.3.3/dart flatbuffers-23.3.3/docs flatbuffers-23.3.3/examples flatbuffers-23.3.3/go flatbuffers-23.3.3/grpc flatbuffers-23.3.3/include flatbuffers-23.3.3/java flatbuffers-23.3.3/js flatbuffers-23.3.3/kotlin flatbuffers-23.3.3/lobster flatbuffers-23.3.3/lua flatbuffers-23.3.3/mjs flatbuffers-23.3.3/net flatbuffers-23.3.3/nim flatbuffers-23.3.3/package.json flatbuffers-23.3.3/php flatbuffers-23.3.3/python flatbuffers-23.3.3/readme.md flatbuffers-23.3.3/reflection flatbuffers-23.3.3/rust flatbuffers-23.3.3/samples flatbuffers-23.3.3/scripts flatbuffers-23.3.3/snap flatbuffers-23.3.3/src flatbuffers-23.3.3/swift flatbuffers-23.3.3/swift.swiftformat flatbuffers-23.3.3/tests flatbuffers-23.3.3/ts flatbuffers-23.3.3/tsconfig.json flatbuffers-23.3.3/tsconfig.mjs.json flatbuffers-23.3.3/typescript.bzl flatbuffers-23.3.3/yarn.lock third_party/flatbuffers/ + tar xf /builddir/build/SOURCES/v2.11.1.tar.gz + rm -rf third_party/pybind11/CMakeLists.txt third_party/pybind11/LICENSE third_party/pybind11/MANIFEST.in third_party/pybind11/README.rst third_party/pybind11/SECURITY.md third_party/pybind11/docs third_party/pybind11/include third_party/pybind11/noxfile.py third_party/pybind11/pybind11 third_party/pybind11/pyproject.toml third_party/pybind11/setup.cfg third_party/pybind11/setup.py third_party/pybind11/tests third_party/pybind11/tools + cp -r pybind11-2.11.1/CMakeLists.txt pybind11-2.11.1/LICENSE pybind11-2.11.1/MANIFEST.in pybind11-2.11.1/README.rst pybind11-2.11.1/SECURITY.md pybind11-2.11.1/docs pybind11-2.11.1/include pybind11-2.11.1/noxfile.py pybind11-2.11.1/pybind11 pybind11-2.11.1/pyproject.toml pybind11-2.11.1/setup.cfg pybind11-2.11.1/setup.py pybind11-2.11.1/tests pybind11-2.11.1/tools third_party/pybind11/ + tar xf /builddir/build/SOURCES/tensorpipe-52791a2.tar.gz + rm -rf third_party/tensorpipe/CMakeLists.txt third_party/tensorpipe/CODE_OF_CONDUCT.md third_party/tensorpipe/CONTRIBUTING.md third_party/tensorpipe/LICENSE.txt third_party/tensorpipe/README.md third_party/tensorpipe/cmake third_party/tensorpipe/docs third_party/tensorpipe/setup.py third_party/tensorpipe/tensorpipe third_party/tensorpipe/third_party + cp -r tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/CMakeLists.txt tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/CODE_OF_CONDUCT.md tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/CONTRIBUTING.md tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/LICENSE.txt tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/README.md tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/cmake tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/docs tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/setup.py tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/tensorpipe tensorpipe-52791a2fd214b2a9dc5759d36725909c1daa7f2e/third_party third_party/tensorpipe/ + tar xf /builddir/build/SOURCES/v1.41.0.tar.gz + rm -rf 'third_party/tensorpipe/third_party/libuv/*' + cp -r libuv-1.41.0/AUTHORS libuv-1.41.0/CMakeLists.txt libuv-1.41.0/CONTRIBUTING.md libuv-1.41.0/ChangeLog libuv-1.41.0/LICENSE libuv-1.41.0/LICENSE-docs libuv-1.41.0/MAINTAINERS.md libuv-1.41.0/Makefile.am libuv-1.41.0/README.md libuv-1.41.0/SUPPORTED_PLATFORMS.md libuv-1.41.0/autogen.sh libuv-1.41.0/configure.ac libuv-1.41.0/docs libuv-1.41.0/img libuv-1.41.0/include libuv-1.41.0/libuv-static.pc.in libuv-1.41.0/libuv.pc.in libuv-1.41.0/m4 libuv-1.41.0/src libuv-1.41.0/test libuv-1.41.0/tools libuv-1.41.0/uv_win_longpath.manifest third_party/tensorpipe/third_party/libuv/ + tar xf /builddir/build/SOURCES/libnop-910b558.tar.gz + rm -rf 'third_party/tensorpipe/third_party/libnop/*' + cp -r libnop-910b55815be16109f04f4180e9adee14fb4ce281/AUTHORS libnop-910b55815be16109f04f4180e9adee14fb4ce281/CONTRIBUTING.md libnop-910b55815be16109f04f4180e9adee14fb4ce281/LICENSE libnop-910b55815be16109f04f4180e9adee14fb4ce281/Makefile libnop-910b55815be16109f04f4180e9adee14fb4ce281/README.md libnop-910b55815be16109f04f4180e9adee14fb4ce281/build libnop-910b55815be16109f04f4180e9adee14fb4ce281/docs libnop-910b55815be16109f04f4180e9adee14fb4ce281/examples libnop-910b55815be16109f04f4180e9adee14fb4ce281/include libnop-910b55815be16109f04f4180e9adee14fb4ce281/test third_party/tensorpipe/third_party/libnop/ + tar xf /builddir/build/SOURCES/v1.14.2.tar.gz + rm -rf third_party/opentelemetry-cpp/CHANGELOG.md third_party/opentelemetry-cpp/CMakeLists.txt third_party/opentelemetry-cpp/CMakeSettings.json third_party/opentelemetry-cpp/CODE_OF_CONDUCT.md third_party/opentelemetry-cpp/CONTRIBUTING.md third_party/opentelemetry-cpp/DEPRECATED.md third_party/opentelemetry-cpp/INSTALL.md third_party/opentelemetry-cpp/LICENSE third_party/opentelemetry-cpp/README.md third_party/opentelemetry-cpp/RELEASING.md third_party/opentelemetry-cpp/Versioning.md third_party/opentelemetry-cpp/WORKSPACE third_party/opentelemetry-cpp/api third_party/opentelemetry-cpp/bazel third_party/opentelemetry-cpp/buildscripts third_party/opentelemetry-cpp/ci third_party/opentelemetry-cpp/cmake third_party/opentelemetry-cpp/docker third_party/opentelemetry-cpp/docs third_party/opentelemetry-cpp/examples third_party/opentelemetry-cpp/exporters third_party/opentelemetry-cpp/ext third_party/opentelemetry-cpp/functional third_party/opentelemetry-cpp/opentracing-shim third_party/opentelemetry-cpp/sdk third_party/opentelemetry-cpp/test_common third_party/opentelemetry-cpp/third_party third_party/opentelemetry-cpp/third_party_release third_party/opentelemetry-cpp/tools + cp -r opentelemetry-cpp-1.14.2/CHANGELOG.md opentelemetry-cpp-1.14.2/CMakeLists.txt opentelemetry-cpp-1.14.2/CMakeSettings.json opentelemetry-cpp-1.14.2/CODE_OF_CONDUCT.md opentelemetry-cpp-1.14.2/CONTRIBUTING.md opentelemetry-cpp-1.14.2/DEPRECATED.md opentelemetry-cpp-1.14.2/INSTALL.md opentelemetry-cpp-1.14.2/LICENSE opentelemetry-cpp-1.14.2/README.md opentelemetry-cpp-1.14.2/RELEASING.md opentelemetry-cpp-1.14.2/Versioning.md opentelemetry-cpp-1.14.2/WORKSPACE opentelemetry-cpp-1.14.2/api opentelemetry-cpp-1.14.2/bazel opentelemetry-cpp-1.14.2/buildscripts opentelemetry-cpp-1.14.2/ci opentelemetry-cpp-1.14.2/cmake opentelemetry-cpp-1.14.2/docker opentelemetry-cpp-1.14.2/docs opentelemetry-cpp-1.14.2/examples opentelemetry-cpp-1.14.2/exporters opentelemetry-cpp-1.14.2/ext opentelemetry-cpp-1.14.2/functional opentelemetry-cpp-1.14.2/opentracing-shim opentelemetry-cpp-1.14.2/sdk opentelemetry-cpp-1.14.2/test_common opentelemetry-cpp-1.14.2/third_party opentelemetry-cpp-1.14.2/third_party_release opentelemetry-cpp-1.14.2/tools third_party/opentelemetry-cpp/ + tar xf /builddir/build/SOURCES/cpp-httplib-3b6597b.tar.gz + rm -rf third_party/cpp-httplib/CMakeLists.txt third_party/cpp-httplib/LICENSE third_party/cpp-httplib/README.md third_party/cpp-httplib/cmake third_party/cpp-httplib/example third_party/cpp-httplib/httplib.h third_party/cpp-httplib/meson.build third_party/cpp-httplib/meson_options.txt third_party/cpp-httplib/split.py third_party/cpp-httplib/test + cp -r cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/CMakeLists.txt cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/LICENSE cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/README.md cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/cmake cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/example cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/httplib.h cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/meson.build cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/meson_options.txt cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/split.py cpp-httplib-3b6597bba913d51161383657829b7e644e59c006/test third_party/cpp-httplib/ + tar xf /builddir/build/SOURCES/kineto-be13176.tar.gz + rm -rf third_party/kineto/CODE_OF_CONDUCT.md third_party/kineto/CONTRIBUTING.md third_party/kineto/LICENSE third_party/kineto/README.md third_party/kineto/libkineto third_party/kineto/tb_plugin + cp -r kineto-be1317644c68b4bfc4646024a6b221066e430031/CODE_OF_CONDUCT.md kineto-be1317644c68b4bfc4646024a6b221066e430031/CONTRIBUTING.md kineto-be1317644c68b4bfc4646024a6b221066e430031/LICENSE kineto-be1317644c68b4bfc4646024a6b221066e430031/README.md kineto-be1317644c68b4bfc4646024a6b221066e430031/libkineto kineto-be1317644c68b4bfc4646024a6b221066e430031/tb_plugin third_party/kineto/ + sed -i -e 's@"gfx90a", "gfx940", "gfx941", "gfx942"@"gfx90a"@' aten/src/ATen/native/cuda/Blas.cpp + sed -i -e /aotriton.cmake/d cmake/Dependencies.cmake + sed -i -e s@fmt::fmt-header-only@fmt@ CMakeLists.txt + sed -i -e s@fmt::fmt-header-only@fmt@ c10/CMakeLists.txt + sed -i -e s@fmt::fmt-header-only@fmt@ torch/CMakeLists.txt + sed -i -e s@fmt::fmt-header-only@fmt@ cmake/Dependencies.cmake + sed -i -e 's@add_subdirectory(${PROJECT_SOURCE_DIR}/third_party/fmt)@#add_subdirectory(${PROJECT_SOURCE_DIR}/third_party/fmt)@' cmake/Dependencies.cmake + sed -i -e 's@set_target_properties(fmt-header-only PROPERTIES INTERFACE_COMPILE_FEATURES "")@#set_target_properties(fmt-header-only PROPERTIES INTERFACE_COMPILE_FEATURES "")@' cmake/Dependencies.cmake + sed -i -e 's@list(APPEND Caffe2_DEPENDENCY_LIBS fmt::fmt-header-only)@#list(APPEND Caffe2_DEPENDENCY_LIBS fmt::fmt-header-only)@' cmake/Dependencies.cmake + sed -i -e 's@if(NOT TARGET fxdiv)@if(MSVC AND USE_XNNPACK)@' caffe2/CMakeLists.txt + sed -i -e 's@TARGET_LINK_LIBRARIES(torch_cpu PRIVATE fxdiv)@#TARGET_LINK_LIBRARIES(torch_cpu PRIVATE fxdiv)@' caffe2/CMakeLists.txt + sed -i -e 's@check_submodules()$@#check_submodules()@' setup.py + mv third_party/miniz-2.1.0 . + mv third_party/build_bundled.py . + mv third_party/flatbuffers . + mv third_party/pybind11 . + mv third_party/tensorpipe . + mv third_party/opentelemetry-cpp . + mv third_party/cpp-httplib . + mv third_party/kineto . + rm -rf third_party/BUCK.oss third_party/BUILD third_party/FP16 third_party/FXdiv third_party/LICENSES_BUNDLED.txt third_party/METADATA.bzl third_party/NNPACK third_party/README.md third_party/VulkanMemoryAllocator third_party/XNNPACK third_party/benchmark third_party/cpp-httplib.BUILD third_party/cpuinfo third_party/cuda.BUILD third_party/cudnn.BUILD third_party/cudnn_frontend third_party/cudnn_frontend.BUILD third_party/cutlass third_party/cutlass.BUILD third_party/eigen third_party/eigen.BUILD third_party/fbgemm third_party/fmt third_party/fmt.BUILD third_party/foxi third_party/foxi.BUILD third_party/gemmlowp third_party/generate-cpuinfo-wrappers.py third_party/generate-xnnpack-wrappers.py third_party/glog.buck.bzl third_party/gloo third_party/gloo.BUILD third_party/googletest third_party/ideep third_party/ideep.BUILD third_party/ittapi third_party/kineto.BUILD third_party/kineto.buck.bzl third_party/mimalloc third_party/mkl-dnn.BUILD third_party/mkl.BUILD third_party/mkl_headers.BUILD third_party/nccl third_party/nlohmann third_party/onnx third_party/onnx.BUILD third_party/opentelemetry-cpp.BUILD third_party/pocketfft third_party/protobuf third_party/psimd third_party/pthreadpool third_party/python-peachpy third_party/sleef third_party/sleef.BUILD third_party/sleef.bzl third_party/substitution.bzl third_party/tensorflow_cuda_bazel_build third_party/tensorpipe.BUILD third_party/valgrind-headers third_party/xnnpack.buck.bzl third_party/xnnpack_src_defs.bzl third_party/xnnpack_wrapper_defs.bzl third_party/xpu.txt + mv build_bundled.py third_party + mv miniz-2.1.0 third_party + mv flatbuffers third_party + mv pybind11 third_party + mv tensorpipe third_party + mv opentelemetry-cpp third_party + mv cpp-httplib third_party + mv kineto third_party + mkdir third_party/pocketfft + mkdir third_party/valgrind-headers + cp /usr/include/valgrind/cachegrind.h /usr/include/valgrind/callgrind.h /usr/include/valgrind/dhat.h /usr/include/valgrind/drd.h /usr/include/valgrind/helgrind.h /usr/include/valgrind/memcheck.h /usr/include/valgrind/valgrind.h third_party/valgrind-headers + sed -i -e 's@DESTINATION ${PYTHON_LIB_REL_PATH}@DESTINATION ${CMAKE_INSTALL_PREFIX}/${PYTHON_LIB_REL_PATH}@' caffe2/CMakeLists.txt + sed -i -e 's@list(APPEND Caffe2_DEPENDENCY_LIBS foxi_loader)@#list(APPEND Caffe2_DEPENDENCY_LIBS foxi_loader)@' cmake/Dependencies.cmake + ./tools/amd_build/build_amd.py third_party/kineto/libkineto/CMakeLists.txt updated /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/script.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/script.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/extension.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/extension.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class_detail.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class_detail.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/abi-check.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/abi-check.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm_windows/libshm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm_windows/libshm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm_windows/core.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm_windows/core.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/err.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/err.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/alloc_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/alloc_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/manager.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/manager.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/libshm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/libshm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/socket.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/socket.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/core.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm/core.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/_VariableFunctions.pyi.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/_VariableFunctions.pyi.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/return_types.pyi.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/return_types.pyi.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/_nn.pyi.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/_nn.pyi.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/__init__.pyi.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/__init__.pyi.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/cpp_prefix.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/cpp_prefix.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/aoti_runtime/interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/aoti_runtime/interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/aoti_runtime/implementation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/aoti_runtime/implementation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/benchmark/utils/timeit_template.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/benchmark/utils/timeit_template.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/benchmark/utils/valgrind_wrapper/compat_bindings.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/benchmark/utils/valgrind_wrapper/compat_bindings.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/benchmark/utils/valgrind_wrapper/timer_callgrind_template.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/benchmark/utils/valgrind_wrapper/timer_callgrind_template.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/data/datapipes/datapipe.pyi.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/data/datapipes/datapipe.pyi.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Layout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Layout.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/QScheme.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/QScheme.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Layout.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Layout.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_headers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_headers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/stub.c -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/stub.c [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/CudaIPCTypes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/CudaIPCTypes.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DynamicTypes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DynamicTypes.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/MemoryFormat.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/MemoryFormat.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DynamicTypes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DynamicTypes.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DataLoader.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DataLoader.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Generator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Generator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Storage.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Storage.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageMethods.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageMethods.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/serialization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/serialization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Event.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Event.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_dimname.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_dimname.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/PyInterpreter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/PyInterpreter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THConcat.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THConcat.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/serialization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/serialization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Generator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Generator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/TypeInfo.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/TypeInfo.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Stream.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Stream.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Device.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Device.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Device.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Device.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DataLoader.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DataLoader.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Dtype.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Dtype.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/MemoryFormat.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/MemoryFormat.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/CudaIPCTypes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/CudaIPCTypes.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageSharing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageSharing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Stream.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Stream.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Exceptions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Exceptions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Storage.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Storage.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageMethods.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageMethods.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Export.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Export.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/itt.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/itt.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/empty.c -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/empty.c [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/copy_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/copy_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Dtype.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Dtype.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/TypeInfo.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/TypeInfo.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/itt_wrapper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/itt_wrapper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageSharing.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageSharing.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Exceptions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Exceptions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Event.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Event.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/itt_wrapper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/itt_wrapper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Module.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/PyInterpreter.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/PyInterpreter.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_dimname.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_dimname.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/QScheme.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/QScheme.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/nccl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/nccl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_comm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_comm.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/utils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Event.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Event.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Graph.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Graph.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/memory_snapshot.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/memory_snapshot.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/CUDAPluggableAllocator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/CUDAPluggableAllocator.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/THCP.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/THCP.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/CUDAPluggableAllocator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/CUDAPluggableAllocator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Stream.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Stream.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/comm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/comm.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_comm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_comm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/nccl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/nccl.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Stream.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Stream.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/device_set.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/device_set.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/memory_snapshot.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/memory_snapshot.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_nccl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_nccl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/comm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/comm.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_nccl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_nccl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Event.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Event.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Module.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/shared/cudnn.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/shared/cudnn.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/shared/nvtx.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/shared/nvtx.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/shared/cudart.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/shared/cudart.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/functorch/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/functorch/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/functorch/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/functorch/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/resource_guard.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/resource_guard.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_log.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_log.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_opt_limit.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_opt_limit.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_opt_limit.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_opt_limit.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_log.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_log.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/cuda/cuda.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/cuda/cuda.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/file_check.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/file_check.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/hooks_for_testing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/hooks_for_testing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/hooks_for_testing.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/hooks_for_testing.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/catch_utils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/catch_utils.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/file_check.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/file_check.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_custom_class.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_custom_class.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/update_graph_executor_opt.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/update_graph_executor_opt.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_custom_class.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_custom_class.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_interpreter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_interpreter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_dict.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_dict.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/update_graph_executor_opt.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/update_graph_executor_opt.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/utf8_decoding_ignore.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/utf8_decoding_ignore.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/utf8_decoding_ignore.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/utf8_decoding_ignore.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_list.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_list.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/module_python.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/module_python.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_dict.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_dict.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_list.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_list.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ivalue.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ivalue.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tree_views.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tree_views.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tree_views.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tree_views.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_resolver.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_resolver.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_info.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_info.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_exception.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_exception.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_preprocess.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_preprocess.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_resolver.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_resolver.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_graph_builder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_graph_builder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_graph_builder.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_graph_builder.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_backend_preprocess.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_backend_preprocess.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_backend_lib.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/xnnpack_backend_lib.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/executor/xnn_executor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/executor/xnn_executor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/serialization/serializer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/serialization/serializer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/serialization/serializer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/serialization/serializer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/compiler/xnn_compiler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/compiler/xnn_compiler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/compiler/xnn_compiler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/xnnpack/compiler/xnn_compiler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/preprocess.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/preprocess.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/context.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/context.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/backend.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/cpp/backend.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLCompiler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLCompiler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLFeatureProvider.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLFeatureProvider.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLModelWrapper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLModelWrapper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLExecutor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLExecutor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLTensorSpec.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/coreml/objc/PTMCoreMLTensorSpec.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_lib.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_lib.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_preprocess.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_preprocess.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/scope.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/scope.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/attributes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/attributes.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/subgraph_matcher.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/subgraph_matcher.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/irparser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/irparser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/node_hashing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/node_hashing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_node_list.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_node_list.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/irparser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/irparser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir_views.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir_views.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/subgraph_matcher.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/subgraph_matcher.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/scope.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/scope.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/attributes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/attributes.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/node_hashing.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/node_hashing.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer_jit.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer_jit.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/callstack_debug_info_serialization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/callstack_debug_info_serialization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/storage_context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/storage_context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/callstack_debug_info_serialization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/callstack_debug_info_serialization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/type_name_uniquer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/type_name_uniquer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/python_print.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/python_print.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/onnx.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/onnx.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer_jit.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer_jit.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/onnx.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/onnx.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/type_name_uniquer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/type_name_uniquer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/mobile_bytecode_generated.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/mobile_bytecode_generated.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_constants.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_constants.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/python_print.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/python_print.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_helpers.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_helpers.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_helpers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_helpers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/LlgaTensorImpl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/LlgaTensorImpl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/prepare_binary.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/prepare_binary.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/kernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/kernel.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/prepare_binary.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/prepare_binary.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/defer_size_check.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/defer_size_check.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/LlgaTensorImpl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/LlgaTensorImpl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/register_interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/register_interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/interface.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/decompose_silu.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/decompose_silu.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/guard_shape.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/guard_shape.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/layout_propagation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/layout_propagation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_fuser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_fuser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/layout_propagation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/layout_propagation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/defer_size_check.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/defer_size_check.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/kernel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/kernel.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_fuser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_fuser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/guard_shape.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/guard_shape.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/operator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/operator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/decompose_silu.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/decompose_silu.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_helper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_helper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_helper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_helper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_rewriter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/onednn/graph_rewriter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/tensor_desc.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/tensor_desc.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_spec.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_spec.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/tensor_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/tensor_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/arg_spec.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/arg_spec.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/partition_desc.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/partition_desc.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fused_kernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fused_kernel.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/resource_strings.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/resource_strings.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/temp_file.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/temp_file.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/block_codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/block_codegen.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/graph_opt.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/graph_opt.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_codegen.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/intrinsic_symbols.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/intrinsic_symbols.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/block_codegen.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/block_codegen.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_codegen.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_visitor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_visitor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_codegen.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_codegen.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/half_support.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/half_support.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_cloner.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_cloner.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_verifier.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_verifier.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/var_substitutor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/var_substitutor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_cloner.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_cloner.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_jit.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_jit.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/exceptions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/exceptions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_visitor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_visitor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/unique_name_manager.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/unique_name_manager.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_core.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_core.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_random.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_random.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest_randomization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest_randomization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_inference.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_inference.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/stmt.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/stmt.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensorexpr_init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensorexpr_init.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_jit.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_jit.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_codegen.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/llvm_codegen.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/graph_opt.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/graph_opt.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_verifier.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_verifier.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_mutator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_mutator.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_inference.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_inference.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensorexpr_init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensorexpr_init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_printer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_printer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_codegen.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/types.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest_randomization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest_randomization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_intrinsics.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_intrinsics.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_core.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_core.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/unique_name_manager.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/unique_name_manager.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/intrinsic_symbols.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/intrinsic_symbols.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_mutator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_mutator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/pointwise.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/pointwise.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/quantization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/quantization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/norm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/norm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/reduction.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/reduction.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/pointwise.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/pointwise.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/operators.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/operators.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/reduction.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/reduction.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/softmax.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/softmax.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/quantization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/quantization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/softmax.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/softmax.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/norm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/norm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/edit_distance.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/edit_distance.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/canonicalize_modified_loop.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/canonicalize_modified_loop.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/inline_loop_condition.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/inline_loop_condition.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/exit_transforms.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/exit_transforms.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser_constants.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser_constants.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/error_report.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/error_report.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_range.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_range.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_ref.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_ref.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/mini_environment.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/mini_environment.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/lexer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/lexer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/error_report.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/error_report.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/edit_distance.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/edit_distance.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/versioned_symbols.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/versioned_symbols.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/name_mangler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/name_mangler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/strtod.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/strtod.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/strtod.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/strtod.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_type_parser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_type_parser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/lexer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/lexer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_range.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_range.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/builtin_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/builtin_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_type_parser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_type_parser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/versioned_symbols.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/versioned_symbols.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/name_mangler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/name_mangler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/builtin_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/builtin_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/canonicalize_modified_loop.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/canonicalize_modified_loop.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/inline_loop_condition.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/inline_loop_condition.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parse_string_literal.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parse_string_literal.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_matching.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_matching.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/exit_transforms.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/exit_transforms.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_matching.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_matching.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/file_format.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/file_format.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_operators.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_operators.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/prim_ops_registery.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/prim_ops_registery.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/interpreter.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/interpreter.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/register_ops_common_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/register_ops_common_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/observer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/observer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/code.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/code.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/promoted_prim_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/promoted_prim_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_bytecode.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_bytecode.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/upgrader_mobile.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/upgrader_mobile.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/type_parser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/type_parser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/profiler_edge.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/profiler_edge.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/upgrader_mobile.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/upgrader_mobile.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/interpreter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/interpreter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/frame.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/frame.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_bytecode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_bytecode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/observer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/observer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/register_ops_common_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/register_ops_common_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_operators.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_operators.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/method.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/method.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/type_parser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/type_parser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/profiler_edge.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/profiler_edge.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_export_common.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_export_common.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/prim_ops_registery.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/prim_ops_registery.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/promoted_prim_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/promoted_prim_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/aot_compiler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/aot_compiler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/aot_compiler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/aot_compiler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/backend.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/backend.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TensorUtils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TensorUtils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/BuildFeatureTracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/BuildFeatureTracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/MobileModelRunner.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/MobileModelRunner.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TracerRunner.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TracerRunner.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TracerRunner.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TracerRunner.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/CustomClassTracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/CustomClassTracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/MobileModelRunner.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/MobileModelRunner.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/KernelDTypeTracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/KernelDTypeTracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/CustomClassTracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/CustomClassTracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/tracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/tracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/OperatorCallTracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/OperatorCallTracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/KernelDTypeTracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/KernelDTypeTracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TensorUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/TensorUtils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/BuildFeatureTracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/BuildFeatureTracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/OperatorCallTracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/model_tracer/OperatorCallTracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/sequential.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/sequential.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/export_data.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/export_data.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/sequential.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/sequential.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/random.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/random.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/export_data.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/export_data.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/random.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/random.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/optim/sgd.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/optim/sgd.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/optim/sgd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/optim/sgd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport_manager.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport_manager.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/runtime_compatibility.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/runtime_compatibility.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/runtime_compatibility.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/runtime_compatibility.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport_manager.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport_manager.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/loop_unrolling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/loop_unrolling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/annotate_warns.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/annotate_warns.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/metal_rewrite.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/metal_rewrite.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_non_tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_non_tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_alias_sensitive.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_alias_sensitive.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/batch_mm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/batch_mm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_inplace_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_inplace_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_fork_wait.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_fork_wait.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_folding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_folding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_linear_bn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_linear_bn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inplace_check.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inplace_check.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/add_if_then_else.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/add_if_then_else.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion_cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion_cuda.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_transpose.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_transpose.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mkldnn_rewrite.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mkldnn_rewrite.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/autocast.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/autocast.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/pass_manager.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/pass_manager.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/check_strict_fusion.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/check_strict_fusion.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/concat_opt.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/concat_opt.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/autocast.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/autocast.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_grad_of.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_grad_of.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_folding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_folding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/specialize_autogradzero.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/specialize_autogradzero.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_graph.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_graph.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/shape_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/shape_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/device_type_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/device_type_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_alias_sensitive.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_alias_sensitive.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_autodiff_subgraphs.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_autodiff_subgraphs.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/restore_mutation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/restore_mutation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/value_refinement_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/value_refinement_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/common_subexpression_elimination.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/common_subexpression_elimination.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/erase_number_types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/erase_number_types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_dict_idioms.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_dict_idioms.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dtype_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dtype_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_cache.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_cache.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/device_type_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/device_type_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_redundant_profiles.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_redundant_profiles.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_concat_linear.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_concat_linear.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_linear.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_linear.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_autodiff_subgraphs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_autodiff_subgraphs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/add_if_then_else.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/add_if_then_else.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/tensorexpr_fuser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/tensorexpr_fuser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/value_refinement_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/value_refinement_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_autodiff_subgraphs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_autodiff_subgraphs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_exceptions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_exceptions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/variadic_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/variadic_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_conv_bn.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_conv_bn.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_expands.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_expands.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_propagation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_propagation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_linear_bn.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_linear_bn.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_ops_to_mkldnn.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_ops_to_mkldnn.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_list_idioms.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_list_idioms.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/concat_opt.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/concat_opt.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_linear.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_linear.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/tensorexpr_fuser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/tensorexpr_fuser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_ops_to_mkldnn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_ops_to_mkldnn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_conv_bn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_conv_bn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/replacement_of_old_operators.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/replacement_of_old_operators.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/vulkan_rewrite.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/vulkan_rewrite.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/common_subexpression_elimination.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/common_subexpression_elimination.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_dropout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_dropout.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_graph_optimizations.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_graph_optimizations.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_functional_graphs.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_functional_graphs.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_relu.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_relu.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_dropout.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_dropout.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/refine_tuple_types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/refine_tuple_types.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_concat_linear.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_concat_linear.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/erase_number_types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/erase_number_types.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/batch_mm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/batch_mm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/pass_manager.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/pass_manager.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_propagation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_propagation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onednn_graph_fuser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onednn_graph_fuser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_list_idioms.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_list_idioms.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_cache.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_cache.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_dict_idioms.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_dict_idioms.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/normalize_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/normalize_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/metal_rewrite.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/metal_rewrite.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_folding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_folding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/specialize_autogradzero.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/specialize_autogradzero.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/refine_tuple_types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/refine_tuple_types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_graph.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_graph.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_rewrite_helper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_rewrite_helper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_redundant_profiles.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_redundant_profiles.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_tuples.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_tuples.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dtype_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dtype_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/hoist_conv_packed_params.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/hoist_conv_packed_params.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fixup_trace_scope_blocks.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fixup_trace_scope_blocks.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_exceptions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_exceptions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/normalize_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/normalize_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/eliminate_no_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/eliminate_no_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_grad_of.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_grad_of.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/hoist_conv_packed_params.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/hoist_conv_packed_params.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_non_tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_non_tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/prepack_folding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/prepack_folding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_transpose.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_transpose.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inplace_check.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inplace_check.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/requires_grad_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/requires_grad_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/subgraph_rewrite.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/subgraph_rewrite.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/annotate_warns.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/annotate_warns.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/check_strict_fusion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/check_strict_fusion.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_functional_graphs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_functional_graphs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/variadic_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/variadic_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fixup_trace_scope_blocks.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fixup_trace_scope_blocks.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_expands.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_expands.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_rewrite_helper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_rewrite_helper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/prepack_folding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/prepack_folding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/subgraph_rewrite.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/subgraph_rewrite.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_tuples.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_tuples.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/vulkan_rewrite.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/vulkan_rewrite.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_inplace_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_inplace_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/loop_unrolling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/loop_unrolling.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/restore_mutation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/restore_mutation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mkldnn_rewrite.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mkldnn_rewrite.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_graph_optimizations.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_graph_optimizations.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/eliminate_no_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/eliminate_no_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_autodiff_subgraphs.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_autodiff_subgraphs.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mobile_optimizer_type.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mobile_optimizer_type.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_relu.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_relu.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/integer_value_refinement.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/integer_value_refinement.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_folding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_folding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/requires_grad_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/requires_grad_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/replacement_of_old_operators.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/replacement_of_old_operators.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/shape_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/shape_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/integer_value_refinement.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/integer_value_refinement.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_fork_wait.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_fork_wait.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_observers.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_observers.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/quantization_type.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/quantization_type.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/quantization_patterns.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/quantization_patterns.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/dedup_module_uses.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/dedup_module_uses.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/fusion_passes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/fusion_passes.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/dedup_module_uses.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/dedup_module_uses.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/register_packed_params.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/register_packed_params.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/quantization_type.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/quantization_type.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/finalize.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/finalize.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/register_packed_params.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/register_packed_params.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/fusion_passes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/fusion_passes.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_quant_dequant.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_quant_dequant.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_observers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_observers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/finalize.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/finalize.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_quant_dequant.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_quant_dequant.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/subgraph_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/subgraph_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/memory_dag.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/memory_dag.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/memory_dag.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/memory_dag.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/op_registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/op_registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/subgraph_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/subgraph_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/optimization_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/optimization_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/optimization_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/optimization_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/op_registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/op_registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/preprocess_for_onnx.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/preprocess_for_onnx.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/list_model_parameters.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/list_model_parameters.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eval_peephole.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eval_peephole.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/deduplicate_initializers.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/deduplicate_initializers.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/scalar_type_analysis.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/scalar_type_analysis.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/remove_inplace_ops_for_onnx.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/remove_inplace_ops_for_onnx.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/cast_all_constant_to_floating.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/cast_all_constant_to_floating.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/prepare_division_for_onnx.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/prepare_division_for_onnx.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/scalar_type_analysis.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/scalar_type_analysis.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/preprocess_for_onnx.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/preprocess_for_onnx.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/list_model_parameters.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/list_model_parameters.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/unpack_quantized_weights.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/unpack_quantized_weights.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eliminate_unused_items.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eliminate_unused_items.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/unpack_quantized_weights.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/unpack_quantized_weights.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eval_peephole.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eval_peephole.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_map.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_map.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/cast_all_constant_to_floating.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/cast_all_constant_to_floating.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eliminate_unused_items.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eliminate_unused_items.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/deduplicate_initializers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/deduplicate_initializers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/peephole.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/peephole.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/naming.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/naming.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/remove_inplace_ops_for_onnx.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/remove_inplace_ops_for_onnx.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_extraction.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_extraction.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_map.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_map.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/naming.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/naming.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_extraction.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_extraction.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_substitution.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_substitution.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/onnx_log.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/onnx_log.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_substitution.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_substitution.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/onnx_log.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/onnx_log.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_fold.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_fold.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/prepare_division_for_onnx.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/prepare_division_for_onnx.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/peephole.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/peephole.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_fold.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_fold.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/common.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/common.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/common.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/common.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_encapsulation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_encapsulation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_encapsulation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_encapsulation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_conversion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_conversion.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_conversion.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_conversion.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/vararg_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/vararg_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/calculate_necessary_args.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/calculate_necessary_args.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/serialized_shape_function_registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/serialized_shape_function_registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/logging.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/logging.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/custom_operator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/custom_operator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/slice_indices_adjust.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/slice_indices_adjust.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_distributed_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_distributed_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_exception.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_exception.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/vararg_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/vararg_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/exception_message.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/exception_message.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/serialized_shape_function_registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/serialized_shape_function_registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops_fulljit.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops_fulljit.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/simple_graph_executor_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/simple_graph_executor_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/slice_indices_adjust.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/slice_indices_adjust.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_iterator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_iterator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/shape_function_registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/shape_function_registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/print_handler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/print_handler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/variable_tensor_list.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/variable_tensor_list.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_special_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_special_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator_options.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator_options.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/autodiff.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/autodiff.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/logging.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/logging.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/simple_graph_executor_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/simple_graph_executor_impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/instruction.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/instruction.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_cuda_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_cuda_ops.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_exception.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_exception.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_trace.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_trace.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/print_handler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/print_handler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_trace.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_trace.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/instruction.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/instruction.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/autodiff.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/autodiff.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_c10_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_c10_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/preprocess_graph.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/preprocess_graph.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/frame.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/frame.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/preprocess_graph.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/preprocess_graph.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/frame.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/frame.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/code_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/code_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/can_emit_inline.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/can_emit_inline.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/fusion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/fusion.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/fusion.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/fusion.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ProcessedNodeInputs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ProcessedNodeInputs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/processed_node_wrapper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/processed_node_wrapper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ProcessedNodeInputs.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ProcessedNodeInputs.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/passes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/passes.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/memory_planner.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/memory_planner.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/generated_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/generated_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/passes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/passes.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/static_method.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/static_method.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/memory_planner.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/memory_planner.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module_save.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module_save.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/version_map.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/version_map.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/version_map.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/version_map.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mtia/Module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mtia/Module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mtia/Module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mtia/Module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cpu/Module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cpu/Module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cpu/Module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cpu/Module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable_hooks.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable_hooks.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_special_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_special_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_anomaly_mode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_anomaly_mode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/jit_decomp_interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/jit_decomp_interface.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/jit_decomp_interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/jit_decomp_interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_saved_variable_hooks.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_saved_variable_hooks.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_legacy_variable.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_legacy_variable.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_python.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_python.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_torch_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_torch_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_meta.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_meta.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_legacy_variable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_legacy_variable.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_saved_variable_hooks.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_saved_variable_hooks.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_hook.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_hook.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable_info.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable_info.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nested_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nested_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/forward_grad.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/forward_grad.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_metadata.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_metadata.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/TraceTypeManual.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/TraceTypeManual.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_anomaly_mode.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_anomaly_mode.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/InferenceMode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/InferenceMode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_torch_functions_manual.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_torch_functions_manual.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_buffer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_buffer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_buffer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_buffer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_linalg_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_linalg_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/forward_grad.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/forward_grad.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/edge.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/edge.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/grad_mode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/grad_mode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_fft_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_fft_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_sparse_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_sparse_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_hook.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_hook.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_enum_tag.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_enum_tag.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/anomaly_mode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/anomaly_mode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nn_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nn_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_engine.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_engine.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/cpp_hook.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/cpp_hook.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_python.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_python.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function_hook.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function_hook.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/anomaly_mode.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/anomaly_mode.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/symbolic.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/symbolic.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nested_functions_manual.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nested_functions_manual.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_engine.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_engine.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/cpp_hook.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/cpp_hook.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_metadata.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_metadata.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeManual.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeManual.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/python_arg_parsing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/python_arg_parsing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/grad_layout_contract.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/grad_layout_contract.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/lambda_post_hook.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/lambda_post_hook.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/wrap_outputs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/wrap_outputs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/warnings.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/warnings.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/warnings.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/warnings.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/error_messages.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/error_messages.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/pybind.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/pybind.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/comm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/comm.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/basic_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/basic_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/comm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/comm.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/basic_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/basic_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mps/Module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mps/Module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mps/Module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mps/Module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/extra_state.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/extra_state.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cache_entry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cache_entry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/extra_state.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/extra_state.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/eval_frame.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/eval_frame.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/python_compiled_autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/python_compiled_autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/compiled_autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/compiled_autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpp_shim.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpp_shim.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/eval_frame.c -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/eval_frame.c [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpython_defs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpython_defs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/python_compiled_autograd.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/python_compiled_autograd.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/debug_macros.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/debug_macros.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpython_defs.c -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpython_defs.c [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpp_shim.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpp_shim.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cache_entry.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cache_entry.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/python_init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/python_init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/events.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/events.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/events.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/events.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/counters.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/counters.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/python_init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/python_init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/counters.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/counters.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_types.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_memoryformats.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_memoryformats.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/schema_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/schema_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_list.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_list.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/disable_torch_function.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/disable_torch_function.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/object_ptr.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/object_ptr.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pythoncapi_compat.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pythoncapi_compat.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark-inl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark-inl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_compat.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_compat.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/cpp_stacktraces.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/cpp_stacktraces.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pybind.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pybind.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_qschemes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_qschemes.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pyobject_preservation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pyobject_preservation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_list.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_list.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/cuda_enabled.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/cuda_enabled.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_torch_function_mode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_torch_function_mode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pybind.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pybind.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/disable_torch_function.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/disable_torch_function.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_numbers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_numbers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/verbose.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/verbose.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/structseq.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/structseq.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_layouts.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_layouts.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/invalid_arguments.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/invalid_arguments.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_scalars.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_scalars.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_symnode.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_symnode.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/byte_order.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/byte_order.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/out_types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/out_types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pycfunction_helpers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pycfunction_helpers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_flatten.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_flatten.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_apply.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_apply.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_new.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_new.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_tuples.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_tuples.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/verbose.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/verbose.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/six.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/six.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_layouts.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_layouts.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pyobject_preservation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pyobject_preservation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_qschemes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_qschemes.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/object_ptr.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/object_ptr.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_symnode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_symnode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/structseq.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/structseq.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/numpy_stub.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/numpy_stub.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_dtypes.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_dtypes.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_numpy.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_numpy.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_apply.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_apply.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_flatten.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_flatten.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/torch_dispatch_mode.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/torch_dispatch_mode.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/cpp_stacktraces.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/cpp_stacktraces.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_numpy.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_numpy.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/byte_order.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/byte_order.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/device_lazy_init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/device_lazy_init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_new.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_new.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_raii.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_raii.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_memoryformats.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_memoryformats.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/schema_info.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/schema_info.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/device_lazy_init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/device_lazy_init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/invalid_arguments.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/invalid_arguments.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_strings.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_strings.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_dtypes.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_dtypes.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_stub.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_stub.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/out_types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/out_types.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_client_interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_client_interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/perf.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/perf.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/data_flow.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/data_flow.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/events.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/events.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/containers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/containers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/perf.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/perf.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/data_flow.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/data_flow.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/perf-inl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/perf-inl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/dwarf_symbolize_enums.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/dwarf_symbolize_enums.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/range_table.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/range_table.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/dwarf_enums.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/dwarf_enums.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/lexer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/lexer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind_error.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind_error.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/fast_symbolizer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/fast_symbolizer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/mem_file.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/mem_file.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/fde.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/fde.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/action.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/action.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwinder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwinder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/sections.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/sections.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/eh_frame_hdr.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/eh_frame_hdr.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/debug_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/debug_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/communicate.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/communicate.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/line_number_program.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/line_number_program.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind_fb.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/unwind/unwind_fb.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/vulkan.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/vulkan.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/vulkan.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/vulkan.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/python_tracer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/python_tracer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/python_tracer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/python_tracer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/combined_traceback.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/combined_traceback.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/pybind.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/pybind.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/combined_traceback.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/combined_traceback.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/base.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/base.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/cuda.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/base.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/itt.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/itt.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/execution_trace_observer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/execution_trace_observer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/privateuse1_observer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/privateuse1_observer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/privateuse1_observer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/privateuse1_observer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/itt_observer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/itt_observer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/execution_trace_observer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/execution_trace_observer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/itt_observer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/itt_observer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/inductor_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/inductor_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/inductor_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/inductor_ops.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/model_container.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/model_container.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/interface.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/thread_local.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/thread_local.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/arrayref_tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/arrayref_tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/utils_cuda.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/utils_cuda.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/scalar_to_tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/scalar_to_tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/model.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/model.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/device_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runtime/device_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_meta_info.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_meta_info.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_meta_info.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_meta_info.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cuda.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/pybind.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/pybind.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/pybind.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/pybind.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cpu.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cpu.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cuda.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cuda.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cpu.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner_cpu.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/model_container_runner.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/tensor_converter.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/tensor_converter.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/mkldnn_tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/mkldnn_tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/proxy_executor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/proxy_executor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/shim_common.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/shim_common.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/tensor_converter.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/tensor_converter.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/shim_cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/shim_cuda.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/mkldnn_tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/mkldnn_tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/c/shim.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/c/shim.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/python_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/python_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/python_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/python_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/tensor_aten_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/tensor_aten_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_lowering_context.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_lowering_context.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/tensor_aten_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/tensor_aten_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ir_builder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ir_builder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node_lowering.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node_lowering.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_autograd_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_autograd_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_lowering_context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_lowering_context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/config.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/config.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_native_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_native_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_eager_fallback.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_eager_fallback.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_autograd_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_autograd_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/dynamic_ir.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/dynamic_ir.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/dynamic_ir.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/dynamic_ir.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/config.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/config.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_eager_fallback.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_eager_fallback.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node_lowering.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node_lowering.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/to_copy.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/to_copy.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/generic.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/generic.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/device_data.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/device_data.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/generic.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/generic.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/device_data.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/device_data.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/thread_pool.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/thread_pool.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/helpers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/helpers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/multi_wait.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/multi_wait.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/multi_wait.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/multi_wait.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_builder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_builder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape_inference.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape_inference.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/lazy_graph_executor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/lazy_graph_executor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/permutation_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/permutation_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/trie.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/trie.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/thread_pool.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/thread_pool.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_dump_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_dump_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_metadata.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_metadata.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/config.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/config.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/hash.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/hash.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape_inference.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape_inference.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_metadata.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_metadata.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/metrics.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/metrics.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/hash.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/hash.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/unique.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/unique.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/helpers.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/helpers.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/cache.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/cache.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/metrics.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/metrics.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_util.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_util.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/permutation_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/permutation_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/dynamic_ir.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/dynamic_ir.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/config.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/config.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_dump_util.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_dump_util.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/lazy_graph_executor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/lazy_graph_executor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/trie.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/trie.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/arithmetic_ir_ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/arithmetic_ir_ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/arithmetic_ir_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/arithmetic_ir_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/internal_ops/ltc_ops.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/internal_ops/ltc_ops.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCUtils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCUtils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/WinSockUtils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/WinSockUtils.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer_cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer_cuda.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/RankLocal.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/RankLocal.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreLibUvBackend.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreLibUvBackend.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UnixSockUtils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UnixSockUtils.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logging.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logging.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCUtils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCUtils.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupMPI.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupMPI.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.hpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/debug.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/debug.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupGloo.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupGloo.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupNCCL.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupNCCL.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupGloo.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupGloo.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logger.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logger.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.cu [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/c10d.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/c10d.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logger.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logger.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/default_comm_hooks.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/default_comm_hooks.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreBackend.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreBackend.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GlooDeviceFactory.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GlooDeviceFactory.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreBackend.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreBackend.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupWrapper.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupWrapper.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/NCCLUtils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/NCCLUtils.hpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer_timer.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer_timer.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/socket.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/socket.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ParamCommsUtils.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ParamCommsUtils.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TraceUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TraceUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FakeProcessGroup.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FakeProcessGroup.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logging.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logging.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCTracing.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCTracing.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PyProcessGroup.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PyProcessGroup.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupCudaP2P.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupCudaP2P.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupUCC.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupUCC.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Work.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Work.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/debug.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/debug.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/NCCLUtils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/NCCLUtils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/error.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/error.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/socket.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/socket.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Ops.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Ops.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupCudaP2P.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupCudaP2P.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Work.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Work.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/HashStore.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/HashStore.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/sequence_num.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/sequence_num.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GlooDeviceFactory.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GlooDeviceFactory.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCTracing.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/UCCTracing.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupWrapper.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupWrapper.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/HashStore.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/HashStore.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupNCCL.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupNCCL.hpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/exception.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/exception.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupUCC.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupUCC.hpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ParamCommsUtils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ParamCommsUtils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupMPI.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupMPI.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/default_comm_hooks.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/default_comm_hooks.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.cu [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Functional.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Functional.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/sequence_num.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/sequence_num.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization_gpu.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization_gpu.cu [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization_gpu.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization_gpu.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/Handlers.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/Handlers.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/Handlers.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/Handlers.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/WorkerServer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/WorkerServer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/WorkerServer.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_plane/WorkerServer.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/StoreCollectives.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/StoreCollectives.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/ControlCollectives.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/ControlCollectives.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/StoreCollectives.hpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/StoreCollectives.hpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/python_autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/python_autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/autograd_metadata.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/autograd_metadata.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/autograd_metadata.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/autograd_metadata.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/engine/dist_engine.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/engine/dist_engine.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/engine/dist_engine.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/engine/dist_engine.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/sendrpc_backward.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/sendrpc_backward.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/sendrpc_backward.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/sendrpc_backward.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_proto.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_proto.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_call.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_call.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_call.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_call.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_resp.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_resp.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_proto.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_proto.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_remote_call.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_remote_call.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_remote_call.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_remote_call.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_call.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_call.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_cuda.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_call.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_call.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_resp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_resp.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/testing.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/testing.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/metrics/registry.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/metrics/registry.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/metrics/RpcMetricsHandler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/metrics/RpcMetricsHandler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/multiprocessing/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/multiprocessing/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/multiprocessing/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/multiprocessing/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/tensor/python_tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/tensor/python_tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/tensor/python_tensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/tensor/python_tensor.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/xpu.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/xpu.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/enum.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/enum.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/mps.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/mps.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/cuda.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/cuda.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/imethod.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/imethod.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/jit.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/jit.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/python/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/python/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adamw.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adamw.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adagrad.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adagrad.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/optimizer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/optimizer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/serialize.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/serialize.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/lbfgs.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/lbfgs.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/rmsprop.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/rmsprop.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/sgd.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/sgd.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adam.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adam.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/schedulers/lr_scheduler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/schedulers/lr_scheduler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/schedulers/step_lr.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/schedulers/step_lr.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/schedulers/reduce_on_plateau_scheduler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/schedulers/reduce_on_plateau_scheduler.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize/output-archive.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize/output-archive.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize/input-archive.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize/input-archive.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/embedding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/embedding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/conv.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/conv.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/linear.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/linear.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/pooling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/pooling.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/padding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/padding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/normalization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/normalization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/adaptive.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/adaptive.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/batchnorm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/batchnorm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/activation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/activation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/instancenorm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/instancenorm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/vision.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/vision.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/dropout.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/dropout.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/transformer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/transformer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/rnn.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/rnn.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/fold.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/fold.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/embedding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/embedding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/upsampling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/upsampling.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/conv.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/conv.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/_functions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/_functions.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/linear.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/linear.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/pixelshuffle.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/pixelshuffle.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/pooling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/pooling.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/padding.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/padding.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/loss.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/loss.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/normalization.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/normalization.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/adaptive.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/adaptive.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/batchnorm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/batchnorm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/distance.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/distance.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/activation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/activation.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/instancenorm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/instancenorm.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/dropout.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/dropout.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/transformer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/transformer.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/rnn.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/rnn.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/container/functional.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/container/functional.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/datasets/mnist.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/datasets/mnist.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/sequential.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/sequential.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/stream.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/stream.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/distributed.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/distributed.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/random.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/random.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/all.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/all.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/enum.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/enum.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/python.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/python.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/xpu.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/xpu.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/sparse.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/sparse.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/ordered_dict.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/ordered_dict.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/imethod.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/imethod.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/jit.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/jit.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/torch.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/torch.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/fft.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/fft.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/mps.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/mps.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/version.h.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/version.h.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/arg.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/arg.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/special.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/special.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/expanding_array.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/expanding_array.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/linalg.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/linalg.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/cuda.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/cuda.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nested.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nested.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/python/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/python/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adamw.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adamw.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adam.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adam.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/lbfgs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/lbfgs.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/optimizer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/optimizer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/sgd.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/sgd.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/rmsprop.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/rmsprop.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/serialize.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/serialize.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adagrad.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adagrad.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/schedulers/reduce_on_plateau_scheduler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/schedulers/reduce_on_plateau_scheduler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/schedulers/step_lr.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/schedulers/step_lr.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/schedulers/lr_scheduler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/schedulers/lr_scheduler.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/TensorDataContainer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/TensorDataContainer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/archive.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/archive.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/input-archive.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/input-archive.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/output-archive.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/output-archive.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/pimpl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/pimpl.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/pimpl-inl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/pimpl-inl.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/parallel/data_parallel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/parallel/data_parallel.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/loss.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/loss.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/linear.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/linear.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/batchnorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/batchnorm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/pooling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/pooling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/fold.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/fold.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/normalization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/normalization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/activation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/activation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/instancenorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/instancenorm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/vision.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/vision.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/upsampling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/upsampling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/dropout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/dropout.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/pixelshuffle.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/pixelshuffle.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/padding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/padding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/embedding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/embedding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/conv.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/conv.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/distance.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/distance.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformercoder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformercoder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/loss.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/loss.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/linear.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/linear.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/batchnorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/batchnorm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/pooling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/pooling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/fold.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/fold.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/normalization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/normalization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformerlayer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformerlayer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/activation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/activation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/rnn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/rnn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/instancenorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/instancenorm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/vision.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/vision.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/upsampling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/upsampling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/dropout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/dropout.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/pixelshuffle.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/pixelshuffle.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/padding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/padding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/embedding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/embedding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/conv.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/conv.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/adaptive.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/adaptive.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/distance.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/distance.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils/clip_grad.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils/clip_grad.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils/convert_parameters.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils/convert_parameters.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils/rnn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/utils/rnn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/transformercoder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/transformercoder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/loss.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/loss.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/linear.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/linear.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/batchnorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/batchnorm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/common.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/common.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/pooling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/pooling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/transformer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/transformer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/fold.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/fold.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/utils.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/normalization.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/normalization.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/transformerlayer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/transformerlayer.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/activation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/activation.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/rnn.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/rnn.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/instancenorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/instancenorm.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/_functions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/_functions.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/upsampling.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/upsampling.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/dropout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/dropout.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/pixelshuffle.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/pixelshuffle.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/padding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/padding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/embedding.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/embedding.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/conv.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/conv.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/adaptive.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/adaptive.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/distance.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/distance.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/parameterdict.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/parameterdict.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/named_any.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/named_any.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/moduledict.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/moduledict.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/parameterlist.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/parameterlist.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/functional.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/functional.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/modulelist.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/modulelist.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/sequential.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/sequential.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/worker_exception.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/worker_exception.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/iterator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/iterator.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader_options.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader_options.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/example.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/example.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/stateful.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/stateful.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/mnist.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/mnist.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/chunk.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/chunk.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/shared.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/shared.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/map.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/map.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/base.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/lambda.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/lambda.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/collate.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/collate.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/stack.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/stack.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/tensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/tensor.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/transforms/base.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader/stateful.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader/stateful.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader/stateless.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader/stateless.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader/base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/dataloader/base.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/detail/queue.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/detail/queue.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/detail/data_shuttle.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/detail/data_shuttle.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/detail/sequencers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/detail/sequencers.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/distributed.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/distributed.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/stream.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/stream.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/sequential.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/sequential.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/random.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/random.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/custom_batch_request.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/custom_batch_request.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/serialize.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/serialize.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/back_compat.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/back_compat.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/init.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/init.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/init.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/init.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/onnx.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/onnx.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/diagnostics/diagnostics.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/diagnostics/diagnostics.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/diagnostics/generated/rules.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/diagnostics/generated/rules.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Event.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Event.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Stream.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Stream.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Stream.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Stream.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Event.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Event.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Module.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Module.h [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Module.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/xpu/Module.cpp [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/nn/functional.pyi.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/nn/functional.pyi.in [skipped, already hipified] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/tools/autograd/templates/python_variable_methods.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/tools/autograd/templates/python_variable_methods.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAException.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPException.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDADeviceAssertion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPDeviceAssertion.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAStream.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPStream.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAStream.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPStream.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAException.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPException.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/driver_api.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/driver_api.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAMallocAsyncAllocator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPMallocAsyncAllocator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAFunctions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPFunctions.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDACachingAllocator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPCachingAllocator.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDADeviceAssertionHost.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPDeviceAssertionHost.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAGraphsC10Utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPGraphsC10Utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAGuard.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPGuard.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAFunctions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPFunctions.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAMiscFunctions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPMiscFunctions.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAMacros.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPMacros.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAAlgorithm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPAlgorithm.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAAllocatorConfig.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPAllocatorConfig.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDADeviceAssertionHost.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPDeviceAssertionHost.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAMiscFunctions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPMiscFunctions.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAAllocatorConfig.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPAllocatorConfig.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDAMathCompat.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPMathCompat.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/driver_api.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/driver_api.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/CUDACachingAllocator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/HIPCachingAllocator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/impl/CUDATest.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/impl/HIPTest.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/impl/cuda_cmake_macros.h.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/impl/hip_cmake_macros.h.in [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/impl/CUDATest.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/impl/HIPTest.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/impl/CUDAGuardImpl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/impl/HIPGuardImpl.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/impl/CUDAGuardImpl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/impl/HIPGuardImpl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/CMakeLists.txt -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/CMakeLists.txt [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_multiple_writes_from_same_block.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_multiple_writes_from_same_block.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_multiple_writes_from_multiple_blocks.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_multiple_writes_from_multiple_blocks.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_from_2_processes.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_from_2_processes.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_catches_thread_and_block_and_device.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_catches_thread_and_block_and_device.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_1_var_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_1_var_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_catches_stream.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_catches_stream.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDATest.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPTest.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/cuda/test/impl/CUDAAssertionsTest_multiple_writes_from_blocks_and_threads.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/hip/test/impl/HIPAssertionsTest_multiple_writes_from_blocks_and_threads.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/PinnedMemoryAllocator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PinnedMemoryAllocator.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAGeneratorImpl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGeneratorImpl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAGraph.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGraph.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/ApplyGridUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/ApplyGridUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAApplyUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAConfig.h.in -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPConfig.h.in [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDASparse.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparse.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDADataType.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPDataType.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/EmptyTensor.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/EmptyTensor.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/cub-RadixSortKeys.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/Atomic.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CachingHostAllocator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CachingHostAllocator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAContextLight.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContextLight.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/ScanUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/ScanUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/cub_definitions.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub_definitions.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/PeerToPeerAccess.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PeerToPeerAccess.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAEvent.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPEvent.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDASparseBlas.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseBlas.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAContext.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/AsmUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/AsmUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/llvm_basic.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/llvm_basic.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/ATenCUDAGeneral.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/ATenHIPGeneral.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDADevice.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPDevice.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CachingHostAllocator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CachingHostAllocator.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/PeerToPeerAccess.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PeerToPeerAccess.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAGeneratorImpl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGeneratorImpl.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/jiterator.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/jiterator.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/ThrustAllocator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/ThrustAllocator.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/PhiloxUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PhiloxUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDASparseDescriptors.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseDescriptors.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDATensorMethods.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPTensorMethods.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/llvm_jit_strings.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/llvm_jit_strings.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/Sleep.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Sleep.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/Exceptions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Exceptions.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/PinnedMemoryAllocator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PinnedMemoryAllocator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/DeviceUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/DeviceUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/NumericLimits.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/NumericLimits.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CuSparseHandlePool.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CuSparseHandlePool.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/llvm_complex.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/llvm_complex.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/cub.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/jiterator_impl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/jiterator_impl.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/cub.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAGraphsUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGraphsUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDABlas.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDABlas.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/jiterator.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/jiterator.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/PhiloxCudaState.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PhiloxCudaState.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDASparseBlas.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseBlas.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/cub.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/cub-RadixSortPairs.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CublasHandlePool.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CublasHandlePool.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/EmptyTensor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/EmptyTensor.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAContext.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/Exceptions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Exceptions.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDAGraph.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGraph.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/Sleep.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Sleep.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/CUDASparseDescriptors.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseDescriptors.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/DeviceThreadHandles.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/DeviceThreadHandles.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/IndexUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/IndexUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/CUDAHooks.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/HIPHooks.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/KernelUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/KernelUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/PhiloxCudaStateRaw.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/PhiloxCudaStateRaw.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/IntegerDivider.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/IntegerDivider.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/LazyNVRTC.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/LazyNVRTC.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/CUDAHooks.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/HIPHooks.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/UnpackRaw.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/UnpackRaw.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/IndexUtils.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/IndexUtils.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/OffsetCalculator.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/OffsetCalculator.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/LazyNVRTC.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/LazyNVRTC.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/detail/TensorInfo.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/TensorInfo.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/GemmRocblas.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/Tunable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/Tunable.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/TunableOp.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableOp.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/TunableGemm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/GemmCommon.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmCommon.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/StreamTimer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/StreamTimer.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/GemmHipblaslt.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmHipblaslt.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/Tunable.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/Tunable.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/tunable/StreamTimer.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/StreamTimer.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/nvrtc_stub/ATenNVRTC.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cuda/nvrtc_stub/ATenNVRTC.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Equal.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Equal.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AdaptiveMaxPooling3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricCosKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricCosKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleBilinear2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/modified_bessel_i1.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/modified_bessel_i1.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachBinaryOpScalarTensor.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachBinaryOpScalarTensor.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/NLLLoss2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LegacyThrustHelpers.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LegacyThrustHelpers.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationThresholdKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationThresholdKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/spherical_bessel_j0.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/spherical_bessel_j0.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricTanKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricTanKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionCauchyKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adam_amsgrad_impl.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adam_amsgrad_impl.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryMiscBackwardOpsKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryMiscBackwardOpsKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceAMinMaxKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceAMinMaxKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceMaxValuesKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceMaxValuesKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/EmbeddingBackwardKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Sort.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sort.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/shifted_chebyshev_polynomial_v.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/shifted_chebyshev_polynomial_v.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Distributions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGammaKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGammaKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationPreluKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationPreluKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryBitwiseOpsKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryBitwiseOpsKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/jit_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/jit_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/IGammaKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IGammaKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/shifted_chebyshev_polynomial_t.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/shifted_chebyshev_polynomial_t.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SortStable.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryRemainderKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryRemainderKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SortStable.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/int4mm.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/int4mm.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationHardswishKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationHardswishKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorModeKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorModeKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/JitLoops.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/JitLoops.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationLeakyReluKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationLeakyReluKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CumminmaxKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumminmaxKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/bessel_j1.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/bessel_j1.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Copy.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Copy.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachMinMaxFunctors.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachMinMaxFunctors.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/NaiveDilatedConvolution.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NaiveDilatedConvolution.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LaunchUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LaunchUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AveragePool2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CuFFTUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CuFFTUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleNearest2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionBernoulli.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SoftMax.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceMomentKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceMomentKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationMishKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationMishKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationHardsigmoidKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationHardsigmoidKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceMinValuesKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceMinValuesKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorCompare.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorCompare.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/chebyshev_polynomial_u.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/chebyshev_polynomial_u.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DilatedMaxPool2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnarySpecialOpsKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnarySpecialOpsKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AdaptiveMaxPooling2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistanceKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistanceKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachFunctors.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachFunctors.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Activation.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Activation.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/shifted_chebyshev_polynomial_u.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/shifted_chebyshev_polynomial_u.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryMulKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryMulKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/chebyshev_polynomial_t.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/chebyshev_polynomial_t.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Repeat.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Repeat.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryDivTruncKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryDivTruncKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MultiLabelMarginCriterion.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MultiLabelMarginCriterion.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnarySignKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnarySignKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MultiMarginLoss.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MultiMarginLoss.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryInternal.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryInternal.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/laguerre_polynomial_l.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/laguerre_polynomial_l.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CompareEQKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CompareEQKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/im2col.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/im2col.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LinearAlgebraStubs.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LinearAlgebraStubs.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TriangularOps.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/KernelUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ScanUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Copy.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Copy.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SpectralOps.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SpectralOps.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FusedAdamWKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FusedAdamWKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FractionalMaxPool3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DepthwiseConv3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorTopK.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ValidateCompressedIndicesKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ValidateCompressedIndicesKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorModeKernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorModeKernel.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/bessel_y1.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/bessel_y1.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Normalization.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Reduce.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Reduce.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricSinKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricSinKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorShapeCUDA.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorShapeHIP.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Sorting.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FractionalMaxPool2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricAsinhKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricAsinhKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CumsumKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CopysignKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CopysignKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Randperm.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Randperm.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorShape.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorShape.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricAcosKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricAcosKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adam_impl.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adam_impl.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationSoftplusKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationSoftplusKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceLogicKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceLogicKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AdaptiveAveragePooling3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryMiscOpsKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryMiscOpsKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorModeKernel.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorModeKernel.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SparseBinaryOpIntersectionKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DeviceSqrt.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DeviceSqrt.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Resize.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Resize.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/EmbeddingBackwardKernel.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/chebyshev_polynomial_w.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/chebyshev_polynomial_w.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Bucketization.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Bucketization.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationHardtanhKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationHardtanhKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleNearest3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MultinomialKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MultinomialKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/PersistentSoftmax.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/PersistentSoftmax.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adamw_impl.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adamw_impl.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/GridSampler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SortingRadixSelect.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Col2Im.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Col2Im.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CompareKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CompareKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationGluKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationGluKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/WeightNorm.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/WeightNorm.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adamw_impl.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adamw_impl.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachUnaryOp.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachUnaryOp.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CUDAScalar.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/HIPScalar.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Dropout.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Dropout.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FusedSgdKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FusedSgdKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/modified_bessel_i0.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/modified_bessel_i0.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FunctionOfAMatrixUtilsKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationSoftshrinkKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationSoftshrinkKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachReduceOp.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachReduceOp.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationSiluKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationSiluKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Math.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Math.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DilatedMaxPool3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/GridSampler.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/bessel_y0.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/bessel_y0.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricAtanhKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricAtanhKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/legendre_polynomial_p.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/legendre_polynomial_p.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Pow.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Pow.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricSinhKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricSinhKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Sort.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sort.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryLogKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryLogKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Im2Col.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Im2Col.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryFractionKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryFractionKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CrossKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CrossKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationGeluKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationGeluKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryShiftOpsKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryShiftOpsKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/RecordStream.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RecordStream.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MaxUnpooling.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MaxUnpooling.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/vol2col.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/vol2col.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceSumProdKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceSumProdKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationLogSigmoidKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationLogSigmoidKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionLogNormalKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/StepKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/StepKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionGeometricKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/hermite_polynomial_he.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/hermite_polynomial_he.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Unique.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Unique.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AbsKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AbsKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/chebyshev_polynomial_v.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/chebyshev_polynomial_v.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleTrilinear3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSample.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Loops.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Loops.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Sort.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sort.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adam_impl.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adam_impl.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Embedding.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachPointwiseOp.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachPointwiseOp.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MemoryAccess.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MemoryAccess.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ComplexKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ComplexKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DepthwiseConv2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/GridSampler.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachBinaryOpScalarList.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachBinaryOpScalarList.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachBinaryOpList.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachBinaryOpList.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricAtanKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricAtanKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Lerp.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Lerp.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorTransformations.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorTopK.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionUniform.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CUDAJitLoops.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/HIPJitLoops.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UniqueCub.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricAsinKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricAsinKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AveragePool3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adam_amsgrad_impl.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adam_amsgrad_impl.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ScatterGatherKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionNormal.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Resize.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Resize.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SortImpl.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortImpl.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CompositeRandomAccessor.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CompositeRandomAccessor.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleNearest1d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Distributions.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/jit_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/jit_utils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryDivFloorKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryDivFloorKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Loss.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Loss.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/NaiveConvolutionTranspose2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NaiveConvolutionTranspose2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Nonzero.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorCompare.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorCompare.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionTemplates.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LogcumsumexpKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/scaled_modified_bessel_k0.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/scaled_modified_bessel_k0.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Distributions.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LinearAlgebra.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LinearAlgebra.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryDivTrueKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryDivTrueKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceOps.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceOps.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryComplexKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryComplexKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Normalization.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorTopK.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/block_reduce.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/block_reduce.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryGeometricKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryGeometricKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceArgMaxKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceArgMaxKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorModeKernel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorModeKernel.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/RreluWithNoise.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ZetaKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ZetaKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachBinaryOpScalar.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachBinaryOpScalar.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MaxMinElementwiseKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MaxMinElementwiseKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FlattenIndicesKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MixedDtypesLinear.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MixedDtypesLinear.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MiscUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MiscUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FusedAdamKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FusedAdamKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ConvolutionMM2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ConvolutionMM2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/PointwiseOpsKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/PointwiseOpsKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleLinear1d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/bessel_j0.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/bessel_j0.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ForeachTernaryOp.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ForeachTernaryOp.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/TensorFactories.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Reduce.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Reduce.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/layer_norm_kernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/layer_norm_kernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryOpsKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryOpsKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReflectionPad.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/IndexKernel.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/group_norm_kernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/group_norm_kernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/BinaryLogicalOpsKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/BinaryLogicalOpsKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SparseMM.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseMM.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionRandomKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceOps.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceOps.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/thread_constants.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/thread_constants.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricAcoshKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricAcoshKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/modified_bessel_k0.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/modified_bessel_k0.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/PowKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/PowKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adam_utils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adam_utils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CUDALoops.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/HIPLoops.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnfoldBackwardKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnfoldBackwardKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/FillKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FillKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/EmbeddingBag.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/NaiveConvolutionTranspose3d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NaiveConvolutionTranspose3d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Shape.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Shape.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/IndexKernel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/modified_bessel_k1.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/modified_bessel_k1.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricCoshKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricCoshKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/airy_ai.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/airy_ai.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/RNN.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/GcdLcmKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GcdLcmKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UpSampleBicubic2d.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationEluKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationEluKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LogAddExpKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogAddExpKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/RangeFactories.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RangeFactories.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adamw_amsgrad_impl.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adamw_amsgrad_impl.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/reduction_template.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/reduction_template.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ActivationHardshrinkKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ActivationHardshrinkKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/scaled_modified_bessel_k1.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/scaled_modified_bessel_k1.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReplicationPadding.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceNormKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceNormKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CumprodKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ScanKernels.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanKernels.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UnaryGeometricTanhKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UnaryGeometricTanhKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/GridSampler.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/MultiTensorApply.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/MultiTensorApply.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Indexing.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/UniqueCub.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/RenormKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RenormKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Randperm.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Randperm.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/DistributionExponentialKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ReduceArgMinKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceArgMinKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Sorting.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/hermite_polynomial_h.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/hermite_polynomial_h.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/CuFFTPlanCache.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CuFFTPlanCache.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SortUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Activation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Activation.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/IndexKernel.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AmpKernels.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AmpKernels.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/ScanKernels.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanKernels.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SpectralOps.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SpectralOps.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/shifted_chebyshev_polynomial_w.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/shifted_chebyshev_polynomial_w.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SummaryOps.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SortingCommon.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingCommon.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Blas.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Blas.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/SegmentReduce.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/Sorting.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/fused_adamw_amsgrad_impl.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/fused_adamw_amsgrad_impl.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/AdaptiveAveragePooling.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/LossCTC.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/epilogue_helpers.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/epilogue_helpers.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/tile_interleaved_layout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/tile_interleaved_layout.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/ft_gemm_configs.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/ft_gemm_configs.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/interleaved_numeric_conversion.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/interleaved_numeric_conversion.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/arch/mma.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/arch/mma.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/warp/mma_tensorop_compute_B_with_f16.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/warp/mma_tensorop_compute_B_with_f16.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/warp/default_mma_tensor_op.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/warp/default_mma_tensor_op.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/warp/mma_tensorop_dequantizer.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/warp/mma_tensorop_dequantizer.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/kernel/default_fpA_intB_traits.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/kernel/default_fpA_intB_traits.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/kernel/mixed_gemm_B_layout.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/kernel/mixed_gemm_B_layout.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/kernel/fpA_intB_gemm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/kernel/fpA_intB_gemm.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/default_mma_bf16.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/default_mma_bf16.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/default_mma.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/default_mma.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/dq_mma_base.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/dq_mma_base.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/dq_mma_pipelined.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/dq_mma_pipelined.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/dq_mma_multistage.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/dq_mma_multistage.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/default_dq_mma_multistage.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/default_dq_mma_multistage.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/default_dq_mma.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/default_dq_mma.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/gemm/threadblock/default_dq_mma_pipelined.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/gemm/threadblock/default_dq_mma_pipelined.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/cutlass_extensions/epilogue/thread/ft_fused_activations.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/cutlass_extensions/epilogue/thread/ft_fused_activations.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebraLibBlas.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLibBlas.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/CusolverDnHandlePool.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/CusolverDnHandlePool.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebraLib.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLib.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/CUDASolver.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/HIPSolver.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/MagmaUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/MagmaUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebra.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebraLib.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLib.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cuda/linalg/CUDASolver.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/HIPSolver.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/sdp_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/sdp_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/attention_backward.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/attention_backward.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/sdp_utils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/sdp_utils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/attention.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/attention.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/mem_eff_attention/gemm_kernel_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/mem_eff_attention/gemm_kernel_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/mem_eff_attention/debug_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/mem_eff_attention/debug_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/mem_eff_attention/pytorch_utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/mem_eff_attention/pytorch_utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/cuda/flash_attn/flash_api.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/flash_attn/flash_api.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/MHA.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/MHA.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/Conv_v7.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/Conv_v7.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/BatchNorm.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/BatchNorm.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/BatchNorm.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/BatchNorm.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/GridSampler.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/GridSampler.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/ConvShared.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/ConvShared.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/ConvShared.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/ConvShared.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/MHA.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/MHA.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/Conv_v8.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/Conv_v8.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/LossCTC.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/LossCTC.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/RNNUtils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/RNNUtils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/AffineGridGenerator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/AffineGridGenerator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/RNN.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/RNN.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/ConvPlaceholders.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/ConvPlaceholders.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/cuda/NestedTensorTransformerFunctions.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorTransformerFunctions.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/cuda/NestedTensorTransformerFunctions.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorTransformerFunctions.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/cuda/NestedTensorMatmul.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorMatmul.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/cuda/NestedTensorBinaryOps.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorBinaryOps.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/cuda/NestedTensorTransformerUtils.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorTransformerUtils.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/Activation.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/Activation.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/IntReprQuant.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/IntReprQuant.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/MakePerTensorQuantizedTensor.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/MakePerTensorQuantizedTensor.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/FusedObsFakeQuant.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/FusedObsFakeQuant.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/AffineQuantizer.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/AffineQuantizer.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/EmbeddingBag.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/EmbeddingBag.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/Activation.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/Activation.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cuda/FakeQuantizeCore.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/FakeQuantizeCore.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/Linear.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/Linear.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/Conv.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/Conv.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/utils.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/utils.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/ConvPrepack.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/ConvPrepack.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/Pooling.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/Pooling.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/BinaryOps.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/BinaryOps.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/LinearUnpackImpl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/LinearUnpackImpl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/ConvUnpackImpl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/ConvUnpackImpl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/LinearPrepack.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/LinearPrepack.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseMatMul.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseMatMul.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseSemiStructuredOps.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseSemiStructuredOps.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseBlasLegacy.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasLegacy.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/StaticSort.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/StaticSort.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseBlasImpl.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasImpl.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseBlas.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlas.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SoftMax.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseBlasImpl.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasImpl.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCUDABlas.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCUDATensor.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCsrTensorMath.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseCsrTensorMath.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseSemiSturcturedApply.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseSemiSturcturedApply.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCUDAApplyUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/ComputeSparseTile.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/ComputeSparseTile.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseSemiStructuredPack.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseSemiStructuredPack.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCUDABlas.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseSemiStructuredLinear.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseSemiStructuredLinear.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseSemiStructuredApplyDense.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseSemiStructuredApplyDense.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCUDATensorMath.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseBlasLegacy.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasLegacy.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseSemiStructuredTile.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseSemiStructuredTile.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/SparseCUDATensorMath.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/cuda/cuSPARSELtOps.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/cuSPARSELtOps.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_allocatorTraceTracker_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_allocatorTraceTracker_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/vulkan_quantized_api_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/vulkan_quantized_api_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/packedtensoraccessor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/packedtensoraccessor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/allocator_clone_test.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/allocator_clone_test.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/tensor_iterator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/tensor_iterator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_optional_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_optional_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/vec_test_all_types.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/vec_test_all_types.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_apply_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_apply_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_dlconvertor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_dlconvertor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cpu_profiling_allocator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/cpu_profiling_allocator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_cudnn_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_cudnn_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/reportMemoryUsage_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/reportMemoryUsage_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_half_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_half_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/thread_init_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/thread_init_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_integer_divider_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_integer_divider_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_allocator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_allocator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/scalar_tensor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/scalar_tensor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/lazy_tensor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/lazy_tensor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/legacy_vmap_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/legacy_vmap_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/test_parallel.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/test_parallel.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/NamedTensor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/NamedTensor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_complex_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_complex_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_tensor_interop_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_tensor_interop_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cpu_rng_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/cpu_rng_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/operators_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/operators_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/vulkan_api_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/vulkan_api_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/undefined_tensor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/undefined_tensor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/mobile_memory_cleanup.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/mobile_memory_cleanup.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/type_ptr_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/type_ptr_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_stream_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_stream_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_atomic_ops_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_atomic_ops_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/reportMemoryUsage.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/reportMemoryUsage.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/extension_backend_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/extension_backend_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/type_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/type_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/test_thread_pool_guard.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/test_thread_pool_guard.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/test_assert.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/test_assert.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/native_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/native_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/memory_overlapping_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/memory_overlapping_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/verify_api_visibility.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/verify_api_visibility.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/StorageUtils_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/StorageUtils_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/scalar_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/scalar_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/reduce_ops_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/reduce_ops_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/half_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/half_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/rng_test.h -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/rng_test.h [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_complex_math_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_complex_math_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/xpu_generator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/xpu_generator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/math_kernel_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/math_kernel_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/memory_format_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/memory_format_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/xla_tensor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/xla_tensor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/xpu_device_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/xpu_device_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/broadcast_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/broadcast_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/Dimname_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/Dimname_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_vectorized_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_vectorized_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_cub_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_cub_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_reportMemoryUsage_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_reportMemoryUsage_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_caching_host_allocator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_caching_host_allocator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/mps_test_print.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/mps_test_print.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/tensor_interop_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/tensor_interop_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_device_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_device_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/dispatch_key_set_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/dispatch_key_set_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_packedtensoraccessor_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_packedtensoraccessor_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cpu_allocator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/cpu_allocator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/stride_properties_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/stride_properties_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cpu_caching_allocator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/cpu_caching_allocator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/xpu_caching_host_allocator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/xpu_caching_host_allocator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/wrapdim_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/wrapdim_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/vec_test_all_types.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/vec_test_all_types.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/Dict_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/Dict_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_generator_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_generator_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/vitals.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/vitals.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/quantized_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/quantized_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/pow_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/pow_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/operator_name_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/operator_name_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cuda_distributions_test.cu -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/hip_distributions_test.hip [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/apply_utils_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/apply_utils_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/mps_test_allocator.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/mps_test_allocator.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/ivalue_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/ivalue_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/xnnpack_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/xnnpack_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/MaybeOwned_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/MaybeOwned_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/ExclusivelyOwned_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/ExclusivelyOwned_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/xpu_event_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/xpu_event_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/dlconvertor_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/dlconvertor_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/basic.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/basic.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/cpu_generator_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/cpu_generator_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/weakref_test.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/weakref_test.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/atest.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/hip/atest.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/test_install/main.cpp -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/test/test_install/hip/main.cpp [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/THC/CMakeLists.txt -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/THH/CMakeLists.txt [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/THC/THCDeviceUtils.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/THH/THHDeviceUtils.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/THC/THCAtomics.cuh -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/THH/THHAtomics.cuh [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/binaries/inspect_gpu.cc -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/binaries/hip/inspect_gpu.cc [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/cpp_wrapper_cpu.py -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/cpp_wrapper_cpu.py [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/cpp_wrapper_cuda.py -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/cpp_wrapper_cuda.py [ok] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/wrapper.py -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_inductor/codegen/wrapper.py [ok] Successfully preprocessed all matching files. + sed -i -e s@rocm-core/rocm_version.h@rocm_version.h@ aten/src/ATen/hip/tunable/TunableGemm.h + sed -i -e s@lib/cmake/hip@lib64/cmake/hip@ cmake/public/LoadHIP.cmake + sed -i -e 's@HIP 1.0@HIP MODULE@' cmake/public/LoadHIP.cmake + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.X4dENx + umask 022 + cd /builddir/build/BUILD/python-torch-2.4.0-build + CFLAGS='-O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CFLAGS + CXXFLAGS='-O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + export CXXFLAGS + FFLAGS='-O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib64/gfortran/modules ' + export FFLAGS + FCFLAGS='-O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -I/usr/lib64/gfortran/modules ' + export FCFLAGS + VALAFLAGS=-g + export VALAFLAGS + RUSTFLAGS='-Copt-level=3 -Cdebuginfo=2 -Ccodegen-units=1 -Cstrip=none -Cforce-frame-pointers=yes -Clink-arg=-specs=/usr/lib/rpm/redhat/redhat-package-notes --cap-lints=warn' + export RUSTFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes ' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd pytorch-v2.4.0 ++ cat /proc/cpuinfo ++ grep -m 1 'cpu cores' ++ awk '{ print $4 }' + COMPILE_JOBS=2 + '[' 2x = x ']' + BUILD_MEM=2 + MEM_KB=0 ++ cat /proc/meminfo ++ grep MemTotal ++ awk '{ print $2 }' + MEM_KB=7958088 ++ eval 'expr 7958088 / 1024' +++ expr 7958088 / 1024 + MEM_MB=7771 ++ eval 'expr 7771 / 1024' +++ expr 7771 / 1024 + MEM_GB=7 ++ eval 'expr 1 + 7 / 2' +++ expr 1 + 7 / 2 + COMPILE_JOBS_MEM=4 + '[' 4 -lt 2 ']' + export MAX_JOBS=2 + MAX_JOBS=2 + export CMAKE_EXE_LINKER_FLAGS=-pie + CMAKE_EXE_LINKER_FLAGS=-pie + export BUILD_CUSTOM_PROTOBUF=OFF + BUILD_CUSTOM_PROTOBUF=OFF + export BUILD_NVFUSER=OFF + BUILD_NVFUSER=OFF + export BUILD_SHARED_LIBS=ON + BUILD_SHARED_LIBS=ON + export BUILD_TEST=OFF + BUILD_TEST=OFF + export CMAKE_BUILD_TYPE=RelWithDebInfo + CMAKE_BUILD_TYPE=RelWithDebInfo + export CMAKE_FIND_PACKAGE_PREFER_CONFIG=ON + CMAKE_FIND_PACKAGE_PREFER_CONFIG=ON + export CAFFE2_LINK_LOCAL_PROTOBUF=OFF + CAFFE2_LINK_LOCAL_PROTOBUF=OFF + export INTERN_BUILD_MOBILE=OFF + INTERN_BUILD_MOBILE=OFF + export USE_DISTRIBUTED=OFF + USE_DISTRIBUTED=OFF + export USE_CUDA=OFF + USE_CUDA=OFF + export USE_FAKELOWP=OFF + USE_FAKELOWP=OFF + export USE_FBGEMM=OFF + USE_FBGEMM=OFF + export USE_FLASH_ATTENTION=OFF + USE_FLASH_ATTENTION=OFF + export USE_GOLD_LINKER=ON + USE_GOLD_LINKER=ON + export USE_GLOO=OFF + USE_GLOO=OFF + export USE_ITT=OFF + USE_ITT=OFF + export USE_KINETO=OFF + USE_KINETO=OFF + export USE_LITE_INTERPRETER_PROFILER=OFF + USE_LITE_INTERPRETER_PROFILER=OFF + export USE_LITE_PROTO=OFF + USE_LITE_PROTO=OFF + export USE_MAGMA=OFF + USE_MAGMA=OFF + export USE_MEM_EFF_ATTENTION=OFF + USE_MEM_EFF_ATTENTION=OFF + export USE_MKLDNN=OFF + USE_MKLDNN=OFF + export USE_MPI=OFF + USE_MPI=OFF + export USE_NCCL=OFF + USE_NCCL=OFF + export USE_NNPACK=OFF + USE_NNPACK=OFF + export USE_NUMPY=ON + USE_NUMPY=ON + export USE_OPENMP=ON + USE_OPENMP=ON + export USE_PYTORCH_QNNPACK=OFF + USE_PYTORCH_QNNPACK=OFF + export USE_ROCM=OFF + USE_ROCM=OFF + export USE_SYSTEM_SLEEF=ON + USE_SYSTEM_SLEEF=ON + export USE_SYSTEM_EIGEN_INSTALL=ON + USE_SYSTEM_EIGEN_INSTALL=ON + export USE_SYSTEM_ONNX=ON + USE_SYSTEM_ONNX=ON + export USE_SYSTEM_PYBIND11=OFF + USE_SYSTEM_PYBIND11=OFF + export USE_SYSTEM_LIBS=OFF + USE_SYSTEM_LIBS=OFF + export USE_TENSORPIPE=OFF + USE_TENSORPIPE=OFF + export USE_XNNPACK=ON + USE_XNNPACK=ON + export USE_XPU=OFF + USE_XPU=OFF + export USE_SYSTEM_PTHREADPOOL=ON + USE_SYSTEM_PTHREADPOOL=ON + export USE_SYSTEM_CPUINFO=ON + USE_SYSTEM_CPUINFO=ON + export USE_SYSTEM_FP16=ON + USE_SYSTEM_FP16=ON + export USE_SYSTEM_FXDIV=ON + USE_SYSTEM_FXDIV=ON + export USE_SYSTEM_PSIMD=ON + USE_SYSTEM_PSIMD=ON + export USE_SYSTEM_XNNPACK=ON + USE_SYSTEM_XNNPACK=ON + export USE_DISTRIBUTED=ON + USE_DISTRIBUTED=ON + export USE_TENSORPIPE=ON + USE_TENSORPIPE=ON + export TP_BUILD_LIBUV=OFF + TP_BUILD_LIBUV=OFF + export USE_MPI=ON + USE_MPI=ON + export USE_ROCM=ON + USE_ROCM=ON + export USE_MAGMA=ON + USE_MAGMA=ON ++ hipconfig -p + export HIP_PATH=/usr + HIP_PATH=/usr ++ hipconfig -R + export ROCM_PATH=/usr + ROCM_PATH=/usr ++ hipconfig -l + export HIP_CLANG_PATH=/usr/lib64/llvm18/bin + HIP_CLANG_PATH=/usr/lib64/llvm18/bin ++ /usr/lib64/llvm18/bin/clang -print-resource-dir + RESOURCE_DIR=/usr/lib64/llvm18/bin/../../../lib/clang/18 + export DEVICE_LIB_PATH=/usr/lib64/llvm18/bin/../../../lib/clang/18/amdgcn/bitcode + DEVICE_LIB_PATH=/usr/lib64/llvm18/bin/../../../lib/clang/18/amdgcn/bitcode + gpu=default + module load rocm/default + local _mlredir=0 + '[' -n '' ']' + case " $@ " in + '[' 0 -eq 0 ']' + _module_raw load rocm/default ++ /usr/bin/tclsh /usr/share/Modules/libexec/modulecmd.tcl bash load rocm/default + eval '__MODULES_LMCONFLICT=rocm/default\&rocm; export __MODULES_LMCONFLICT; ROCM_BIN=/usr/bin; export ROCM_BIN; _LMFILES_=/usr/share/modulefiles/rocm/default; export _LMFILES_; LOADEDMODULES=rocm/default; export LOADEDMODULES; ROCM_GPUS=gfx1010\;gfx1012\;gfx1030\;gfx1031\;gfx1100\;gfx1101\;gfx1102; export ROCM_GPUS; ROCM_LIB=/usr/lib64; export ROCM_LIB; test 0;' ++ __MODULES_LMCONFLICT='rocm/default&rocm' ++ export __MODULES_LMCONFLICT ++ ROCM_BIN=/usr/bin ++ export ROCM_BIN ++ _LMFILES_=/usr/share/modulefiles/rocm/default ++ export _LMFILES_ ++ LOADEDMODULES=rocm/default ++ export LOADEDMODULES ++ ROCM_GPUS='gfx1010;gfx1012;gfx1030;gfx1031;gfx1100;gfx1101;gfx1102' ++ export ROCM_GPUS ++ ROCM_LIB=/usr/lib64 ++ export ROCM_LIB ++ test 0 + _mlstatus=0 + return 0 + export 'PYTORCH_ROCM_ARCH=gfx1010;gfx1012;gfx1030;gfx1031;gfx1100;gfx1101;gfx1102' + PYTORCH_ROCM_ARCH='gfx1010;gfx1012;gfx1030;gfx1031;gfx1100;gfx1101;gfx1102' + CFLAGS='-O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes ' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -sP' -- The CXX compiler identification is GNU 14.2.1 -- The C compiler identification is GNU 14.2.1 -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/g++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/gcc - skipped -- Detecting C compile features -- Detecting C compile features - done -- /usr/bin/g++ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/abi-check.cpp -o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/abi-check -- Determined _GLIBCXX_USE_CXX11_ABI=1 -- Not forcing any particular BLAS to be found -- Could not find ccache. Consider installing ccache to speed up compilation. -- Performing Test C_HAS_AVX_1 -- Performing Test C_HAS_AVX_1 - Failed -- Performing Test C_HAS_AVX_2 -- Performing Test C_HAS_AVX_2 - Success -- Performing Test C_HAS_AVX2_1 -- Performing Test C_HAS_AVX2_1 - Failed -- Performing Test C_HAS_AVX2_2 -- Performing Test C_HAS_AVX2_2 - Success -- Performing Test C_HAS_AVX512_1 -- Performing Test C_HAS_AVX512_1 - Failed -- Performing Test C_HAS_AVX512_2 -- Performing Test C_HAS_AVX512_2 - Success -- Performing Test CXX_HAS_AVX_1 -- Performing Test CXX_HAS_AVX_1 - Failed -- Performing Test CXX_HAS_AVX_2 -- Performing Test CXX_HAS_AVX_2 - Success -- Performing Test CXX_HAS_AVX2_1 -- Performing Test CXX_HAS_AVX2_1 - Failed -- Performing Test CXX_HAS_AVX2_2 -- Performing Test CXX_HAS_AVX2_2 - Success -- Performing Test CXX_HAS_AVX512_1 -- Performing Test CXX_HAS_AVX512_1 - Failed -- Performing Test CXX_HAS_AVX512_2 -- Performing Test CXX_HAS_AVX512_2 - Success -- Current compiler supports avx2 extension. Will build perfkernels. -- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX512_EXTENSIONS -- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX512_EXTENSIONS - Success -- Current compiler supports avx512f extension. Will build fbgemm. -- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY -- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY - Success -- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY -- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY - Success -- Performing Test COMPILER_SUPPORTS_RDYNAMIC -- Performing Test COMPILER_SUPPORTS_RDYNAMIC - Success -- Caffe2: Found protobuf with new-style protobuf targets. -- Caffe2 protobuf include directory: /usr/include -- Found Threads: TRUE -- Trying to find preferred BLAS backend of choice: MKL -- MKL_THREADING = OMP -- Looking for sys/types.h -- Looking for sys/types.h - found -- Looking for stdint.h -- Looking for stdint.h - found -- Looking for stddef.h -- Looking for stddef.h - found -- Check size of void* -- Check size of void* - done -- MKL_THREADING = OMP CMake Warning at cmake/Dependencies.cmake:201 (message): MKL could not be found. Defaulting to Eigen Call Stack (most recent call first): CMakeLists.txt:854 (include) CMake Warning at cmake/Dependencies.cmake:243 (message): Preferred BLAS (MKL) cannot be found, now searching for a general BLAS library Call Stack (most recent call first): CMakeLists.txt:854 (include) -- MKL_THREADING = OMP -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl - guide - pthread - m] -- Library mkl: not found -- MKL library not found -- Checking for [blis] -- Library blis: BLAS_blis_LIBRARY-NOTFOUND -- Checking for [Accelerate] -- Library Accelerate: BLAS_Accelerate_LIBRARY-NOTFOUND -- Checking for [vecLib] -- Library vecLib: BLAS_vecLib_LIBRARY-NOTFOUND -- Checking for [flexiblas] -- Library flexiblas: /usr/lib64/libflexiblas.so -- Looking for sgemm_ -- Looking for sgemm_ - found -- Performing Test BLAS_F2C_DOUBLE_WORKS -- Performing Test BLAS_F2C_DOUBLE_WORKS - Failed -- Performing Test BLAS_F2C_FLOAT_WORKS -- Performing Test BLAS_F2C_FLOAT_WORKS - Success -- Performing Test BLAS_USE_CBLAS_DOT -- Performing Test BLAS_USE_CBLAS_DOT - Success -- Looking for sbgemm_ -- Looking for sbgemm_ - not found -- Found a library with BLAS API (flexi). Full path: (/usr/lib64/libflexiblas.so) -- Using pocketfft in directory: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/pocketfft/ -- Found pthreadpool: /usr/lib64/libpthreadpool.so Found cpuinfo: /usr/lib64/libcpuinfo.so -- Found XNNPACK: /usr/lib64/libXNNPACK.so CMake Warning at cmake/Dependencies.cmake:760 (message): Turning USE_FAKELOWP off as it depends on USE_FBGEMM. Call Stack (most recent call first): CMakeLists.txt:854 (include) -- Found Numa: /usr/include -- Found Numa (include: /usr/include, library: /usr/lib64/libnuma.so) -- Found system Eigen at /usr/include/eigen3 -- Found Python: /usr/bin/python3 (found version "3.13.1") found components: Interpreter Development.Module NumPy -- Using third_party/pybind11. -- pybind11 include dirs: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/cmake/../third_party/pybind11/include -- Could NOT find OpenTelemetryApi (missing: OpenTelemetryApi_INCLUDE_DIRS) -- Using third_party/opentelemetry-cpp. -- opentelemetry api include dirs: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/cmake/../third_party/opentelemetry-cpp/api/include -- Checking for module 'mpi-c' -- Package 'mpi-c' not found -- Could NOT find MPI_C (missing: MPI_C_LIB_NAMES MPI_C_HEADER_DIR MPI_C_WORKS) -- Checking for module 'mpi-cxx' -- Package 'mpi-cxx' not found -- Could NOT find MPI_CXX (missing: MPI_CXX_LIB_NAMES MPI_CXX_HEADER_DIR MPI_CXX_WORKS) -- Could NOT find MPI (missing: MPI_C_FOUND MPI_CXX_FOUND) CMake Warning at cmake/Dependencies.cmake:945 (message): Not compiling with MPI. Suppress this warning with -DUSE_MPI=OFF Call Stack (most recent call first): CMakeLists.txt:854 (include) -- MKL_THREADING = OMP -- Check OMP with lib /usr/lib64/libomp.so and flags -fopenmp -v -- MKL_THREADING = OMP -- Check OMP with lib /usr/lib64/libomp.so and flags -fopenmp -v -- Found OpenMP_C: -fopenmp (found version "4.5") -- Found OpenMP_CXX: -fopenmp (found version "4.5") -- Found OpenMP: TRUE (found version "4.5") -- Adding OpenMP CXX_FLAGS: -fopenmp -- Will link against OpenMP libraries: /usr/lib64/libomp.so Building PyTorch for GPU arch: gfx1010;gfx1012;gfx1030;gfx1031;gfx1100;gfx1101;gfx1102 -- Found HIP: /usr (found version "6.2.41134-0") HIP VERSION: 6.2.41134-0 -- Caffe2: Header version is: 6.2.0 ***** ROCm version from rocm_version.h **** ROCM_VERSION_DEV: 6.2.0 ROCM_VERSION_DEV_MAJOR: 6 ROCM_VERSION_DEV_MINOR: 2 ROCM_VERSION_DEV_PATCH: 0 ROCM_VERSION_DEV_INT: 60200 HIP_VERSION_MAJOR: 6 HIP_VERSION_MINOR: 2 TORCH_HIP_VERSION: 602 ***** Library versions from dpkg ***** ***** Library versions from cmake find_package ***** hip VERSION: 6.2.41134 hsa-runtime64 VERSION: 1.14.0 amd_comgr VERSION: 2.8.0 rocrand VERSION: 3.1.0 hiprand VERSION: 2.11.0 rocblas VERSION: 4.2.1 hipblas VERSION: 2.2.0 hipblaslt VERSION: 0.8.0 miopen VERSION: 3.2.0 hipfft VERSION: 1.0.15 hipsparse VERSION: 3.1.1 CMake Warning at cmake/public/LoadHIP.cmake:36 (find_package): By not providing "Findrccl.cmake" in CMAKE_MODULE_PATH this project has asked CMake to find a package configuration file provided by "rccl", but CMake did not find one. Could not find a package configuration file provided by "rccl" with any of the following names: rcclConfig.cmake rccl-config.cmake Add the installation prefix of "rccl" to CMAKE_PREFIX_PATH or set "rccl_DIR" to a directory containing one of the above files. If "rccl" provides a separate development package or SDK, be sure it has been installed. Call Stack (most recent call first): cmake/public/LoadHIP.cmake:162 (find_package_and_print_version) cmake/Dependencies.cmake:1031 (include) CMakeLists.txt:854 (include) rccl VERSION: rocprim VERSION: 3.2.0 hipcub VERSION: 3.2.0 rocthrust VERSION: 3.1.0 hipsolver VERSION: 2.2.0 HIP is using new type enums INFOCompiling with HIP for AMD. -- Disabling Kernel Assert for ROCm TORCH_HIP_VERSION=602 is added as a compiler defines CMake Deprecation Warning at third_party/tensorpipe/third_party/libuv/CMakeLists.txt:1 (cmake_minimum_required): Compatibility with CMake < 3.5 will be removed from a future version of CMake. Update the VERSION argument value or use a ... suffix to tell CMake that the project does not need compatibility with older versions. -- Performing Test UV_LINT_W4 -- Performing Test UV_LINT_W4 - Failed -- Performing Test UV_LINT_NO_UNUSED_PARAMETER_MSVC -- Performing Test UV_LINT_NO_UNUSED_PARAMETER_MSVC - Failed -- Performing Test UV_LINT_NO_CONDITIONAL_CONSTANT_MSVC -- Performing Test UV_LINT_NO_CONDITIONAL_CONSTANT_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_EMPTY_TU_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_EMPTY_TU_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_FILE_SCOPE_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_FILE_SCOPE_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_NONSTATIC_DLIMPORT_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_NONSTATIC_DLIMPORT_MSVC - Failed -- Performing Test UV_LINT_NO_HIDES_LOCAL -- Performing Test UV_LINT_NO_HIDES_LOCAL - Failed -- Performing Test UV_LINT_NO_HIDES_PARAM -- Performing Test UV_LINT_NO_HIDES_PARAM - Failed -- Performing Test UV_LINT_NO_HIDES_GLOBAL -- Performing Test UV_LINT_NO_HIDES_GLOBAL - Failed -- Performing Test UV_LINT_NO_CONDITIONAL_ASSIGNMENT_MSVC -- Performing Test UV_LINT_NO_CONDITIONAL_ASSIGNMENT_MSVC - Failed -- Performing Test UV_LINT_NO_UNSAFE_MSVC -- Performing Test UV_LINT_NO_UNSAFE_MSVC - Failed -- Performing Test UV_LINT_WALL -- Performing Test UV_LINT_WALL - Success -- Performing Test UV_LINT_NO_UNUSED_PARAMETER -- Performing Test UV_LINT_NO_UNUSED_PARAMETER - Success -- Performing Test UV_LINT_STRICT_PROTOTYPES -- Performing Test UV_LINT_STRICT_PROTOTYPES - Success -- Performing Test UV_LINT_EXTRA -- Performing Test UV_LINT_EXTRA - Success -- Performing Test UV_LINT_UTF8_MSVC -- Performing Test UV_LINT_UTF8_MSVC - Failed -- Performing Test UV_F_STRICT_ALIASING -- Performing Test UV_F_STRICT_ALIASING - Success -- summary of build options: Install prefix: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch Target system: Linux Compiler: C compiler: /usr/bin/gcc CFLAGS: -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -- Found uv: 1.38.1 (found version "1.38.1") CMake Warning at cmake/Dependencies.cmake:1168 (message): TensorPipe doesn't yet support ROCm Call Stack (most recent call first): CMakeLists.txt:854 (include) -- Found onnx: /usr/lib64/libonnx.so /usr/lib64/libonnx_proto.so -- Found CUDA with FP16 support, compiling with torch.cuda.HalfTensor -- Removing -DNDEBUG from compile flags -- Checking prototype magma_get_sgeqrf_nb for MAGMA_V2 -- Checking prototype magma_get_sgeqrf_nb for MAGMA_V2 - False -- Compiling with MAGMA support -- MAGMA INCLUDE DIRECTORIES: /usr/include -- MAGMA LIBRARIES: /usr/lib64/libmagma.so -- MAGMA V2 check: 0 CMake Warning (dev) at cmake/Modules/FindARM.cmake:5 (EXEC_PROGRAM): Policy CMP0153 is not set: The exec_program command should not be called. Run "cmake --help-policy CMP0153" for policy details. Use the cmake_policy command to set the policy and suppress this warning. Use execute_process() instead. Call Stack (most recent call first): cmake/Dependencies.cmake:1416 (find_package) CMakeLists.txt:854 (include) This warning is for project developers. Use -Wno-dev to suppress it. -- Could not find hardware support for NEON on this machine. -- No OMAP3 processor on this machine. -- No OMAP4 processor on this machine. -- MKL_THREADING = OMP -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl - guide - pthread - m] -- Library mkl: not found -- MKL library not found -- Checking for [blis] -- Library blis: BLAS_blis_LIBRARY-NOTFOUND -- Checking for [Accelerate] -- Library Accelerate: BLAS_Accelerate_LIBRARY-NOTFOUND -- Checking for [vecLib] -- Library vecLib: BLAS_vecLib_LIBRARY-NOTFOUND -- Checking for [flexiblas] -- Library flexiblas: /usr/lib64/libflexiblas.so -- Found a library with BLAS API (flexi). Full path: (/usr/lib64/libflexiblas.so) -- Looking for cheev_ -- Looking for cheev_ - found -- Found a library with LAPACK API (flexi). disabling CUDA because NOT USE_CUDA is set disabling MKLDNN because USE_MKLDNN is not set -- Looking for clock_gettime in rt -- Looking for clock_gettime in rt - found -- Looking for mmap -- Looking for mmap - found -- Looking for shm_open -- Looking for shm_open - found -- Looking for shm_unlink -- Looking for shm_unlink - found -- Looking for malloc_usable_size -- Looking for malloc_usable_size - found -- -- Performing Test COMPILE_OUT_ZVECTOR -- Performing Test COMPILE_OUT_ZVECTOR - Failed -- ZVECTOR flags were NOT set. -- -- GCC 14.2.1: Adding gcc and gcc_s libs to link line -- Performing Test HAS_WERROR_RETURN_TYPE -- Performing Test HAS_WERROR_RETURN_TYPE - Success -- Performing Test HAS_WERROR_NON_VIRTUAL_DTOR -- Performing Test HAS_WERROR_NON_VIRTUAL_DTOR - Success -- Performing Test HAS_WERROR_BRACED_SCALAR_INIT -- Performing Test HAS_WERROR_BRACED_SCALAR_INIT - Failed -- Performing Test HAS_WERROR_RANGE_LOOP_CONSTRUCT -- Performing Test HAS_WERROR_RANGE_LOOP_CONSTRUCT - Success -- Performing Test HAS_WERROR_BOOL_OPERATION -- Performing Test HAS_WERROR_BOOL_OPERATION - Success -- Performing Test HAS_WNARROWING -- Performing Test HAS_WNARROWING - Success -- Performing Test HAS_WNO_MISSING_FIELD_INITIALIZERS -- Performing Test HAS_WNO_MISSING_FIELD_INITIALIZERS - Success -- Performing Test HAS_WNO_TYPE_LIMITS -- Performing Test HAS_WNO_TYPE_LIMITS - Success -- Performing Test HAS_WNO_ARRAY_BOUNDS -- Performing Test HAS_WNO_ARRAY_BOUNDS - Success -- Performing Test HAS_WNO_UNKNOWN_PRAGMAS -- Performing Test HAS_WNO_UNKNOWN_PRAGMAS - Success -- Performing Test HAS_WNO_UNUSED_PARAMETER -- Performing Test HAS_WNO_UNUSED_PARAMETER - Success -- Performing Test HAS_WNO_UNUSED_FUNCTION -- Performing Test HAS_WNO_UNUSED_FUNCTION - Success -- Performing Test HAS_WNO_UNUSED_RESULT -- Performing Test HAS_WNO_UNUSED_RESULT - Success -- Performing Test HAS_WNO_STRICT_OVERFLOW -- Performing Test HAS_WNO_STRICT_OVERFLOW - Success -- Performing Test HAS_WNO_STRICT_ALIASING -- Performing Test HAS_WNO_STRICT_ALIASING - Success -- Performing Test HAS_WNO_STRINGOP_OVERFLOW -- Performing Test HAS_WNO_STRINGOP_OVERFLOW - Success -- Performing Test HAS_WVLA_EXTENSION -- Performing Test HAS_WVLA_EXTENSION - Failed -- Performing Test HAS_WSUGGEST_OVERRIDE -- Performing Test HAS_WSUGGEST_OVERRIDE - Success -- Performing Test HAS_WNEWLINE_EOF -- Performing Test HAS_WNEWLINE_EOF - Failed -- Performing Test HAS_WINCONSISTENT_MISSING_OVERRIDE -- Performing Test HAS_WINCONSISTENT_MISSING_OVERRIDE - Failed -- Performing Test HAS_WINCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE -- Performing Test HAS_WINCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE - Failed CMake Warning at CMakeLists.txt:1020 (message): USE_GOLD_LINKER was set but ld.gold isn't available, turning it off -- Performing Test HAS_WNO_ERROR_PEDANTIC -- Performing Test HAS_WNO_ERROR_PEDANTIC - Success -- Performing Test HAS_WNO_ERROR_OLD_STYLE_CAST -- Performing Test HAS_WNO_ERROR_OLD_STYLE_CAST - Success -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_OVERRIDE -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_OVERRIDE - Failed -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE - Failed -- Performing Test HAS_WCONSTANT_CONVERSION -- Performing Test HAS_WCONSTANT_CONVERSION - Failed -- Performing Test HAS_WNO_INVALID_PARTIAL_SPECIALIZATION -- Performing Test HAS_WNO_INVALID_PARTIAL_SPECIALIZATION - Failed -- Performing Test HAS_WNO_ALIGNED_ALLOCATION_UNAVAILABLE -- Performing Test HAS_WNO_ALIGNED_ALLOCATION_UNAVAILABLE - Failed -- Performing Test HAS_WNO_MISSING_BRACES -- Performing Test HAS_WNO_MISSING_BRACES - Success -- Performing Test HAS_QUNUSED_ARGUMENTS -- Performing Test HAS_QUNUSED_ARGUMENTS - Failed -- Performing Test HAS_FDIAGNOSTICS_COLOR_ALWAYS -- Performing Test HAS_FDIAGNOSTICS_COLOR_ALWAYS - Success -- Performing Test HAS_FALIGNED_NEW -- Performing Test HAS_FALIGNED_NEW - Success -- Performing Test HAS_WNO_UNUSED_BUT_SET_VARIABLE -- Performing Test HAS_WNO_UNUSED_BUT_SET_VARIABLE - Success -- Performing Test HAS_WNO_MAYBE_UNINITIALIZED -- Performing Test HAS_WNO_MAYBE_UNINITIALIZED - Success -- Performing Test HAS_FSTANDALONE_DEBUG -- Performing Test HAS_FSTANDALONE_DEBUG - Failed -- Performing Test HAS_FNO_MATH_ERRNO -- Performing Test HAS_FNO_MATH_ERRNO - Success -- Performing Test HAS_FNO_TRAPPING_MATH -- Performing Test HAS_FNO_TRAPPING_MATH - Success -- Performing Test HAS_WERROR_FORMAT -- Performing Test HAS_WERROR_FORMAT - Success -- Performing Test HAS_WDEPRECATED -- Performing Test HAS_WDEPRECATED - Success -- NUMA paths: -- /usr/include -- /usr/lib64/libnuma.so -- Looking for backtrace -- Looking for backtrace - found -- backtrace facility detected in default set of libraries -- Found Backtrace: /usr/include -- headers outputs: -- sources outputs: -- declarations_yaml outputs: -- Performing Test COMPILER_SUPPORTS_NO_AVX256_SPLIT -- Performing Test COMPILER_SUPPORTS_NO_AVX256_SPLIT - Success -- Using ATen parallel backend: OMP Building PyTorch for GPU arch: gfx1010;gfx1012;gfx1030;gfx1031;gfx1100;gfx1101;gfx1102 HIP VERSION: 6.2.41134-0 -- Caffe2: Header version is: 6.2.0 ***** ROCm version from rocm_version.h **** ROCM_VERSION_DEV: 6.2.0 ROCM_VERSION_DEV_MAJOR: 6 ROCM_VERSION_DEV_MINOR: 2 ROCM_VERSION_DEV_PATCH: 0 ROCM_VERSION_DEV_INT: 60200 HIP_VERSION_MAJOR: 6 HIP_VERSION_MINOR: 2 TORCH_HIP_VERSION: 602 ***** Library versions from dpkg ***** ***** Library versions from cmake find_package ***** hip VERSION: 6.2.41134 hsa-runtime64 VERSION: 1.14.0 amd_comgr VERSION: 2.8.0 rocrand VERSION: 3.1.0 hiprand VERSION: 2.11.0 rocblas VERSION: 4.2.1 hipblas VERSION: 2.2.0 hipblaslt VERSION: 0.8.0 miopen VERSION: 3.2.0 hipfft VERSION: 1.0.15 hipsparse VERSION: 3.1.1 CMake Warning at cmake/public/LoadHIP.cmake:36 (find_package): By not providing "Findrccl.cmake" in CMAKE_MODULE_PATH this project has asked CMake to find a package configuration file provided by "rccl", but CMake did not find one. Could not find a package configuration file provided by "rccl" with any of the following names: rcclConfig.cmake rccl-config.cmake Add the installation prefix of "rccl" to CMAKE_PREFIX_PATH or set "rccl_DIR" to a directory containing one of the above files. If "rccl" provides a separate development package or SDK, be sure it has been installed. Call Stack (most recent call first): cmake/public/LoadHIP.cmake:162 (find_package_and_print_version) aten/CMakeLists.txt:67 (include) rccl VERSION: rocprim VERSION: 3.2.0 hipcub VERSION: 3.2.0 rocthrust VERSION: 3.1.0 hipsolver VERSION: 2.2.0 HIP is using new type enums ROCm is enabled. Found sleef: /usr/lib64/libsleef.so AT_INSTALL_INCLUDE_DIR include/ATen/core core header install: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h core header install: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/aten_interned_strings.h core header install: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/enum_tag.h -- _GLIBCXX_USE_CXX11_ABI=1 is already defined as a cmake variable INFOcaffe2 ROCM_SOURCE_DIR = -- Using /usr/lib/python3.13/site-packages as python relative installation path CMake Warning at CMakeLists.txt:1283 (message): Generated cmake files are only fully tested if one builds with system glog, gflags, and protobuf. Other settings may generate files that are not well tested. -- -- ******** Summary ******** -- General: -- CMake version : 3.30.5 -- CMake command : /usr/bin/cmake -- System : Linux -- C++ compiler : /usr/bin/g++ -- C++ compiler id : GNU -- C++ compiler version : 14.2.1 -- Using ccache if found : ON -- Found ccache : CCACHE_PROGRAM-NOTFOUND -- CXX flags : -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -D_GLIBCXX_USE_CXX11_ABI=1 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=pedantic -Wno-error=old-style-cast -Wno-missing-braces -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow -- Shared LD flags : -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -Wl,--no-as-needed -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -rdynamic -- Static LD flags : -- Module LD flags : -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -- Build type : RelWithDebInfo -- Compile definitions : ROCM_VERSION=60200;TORCH_HIP_VERSION=602;ONNX_ML=1;ONNXIFI_ENABLE_EXT=1;ONNX_NAMESPACE=onnx;HAVE_MMAP=1;_FILE_OFFSET_BITS=64;HAVE_SHM_OPEN=1;HAVE_SHM_UNLINK=1;HAVE_MALLOC_USABLE_SIZE=1;USE_EXTERNAL_MZCRC;MINIZ_DISABLE_ZIP_READER_CRC32_CHECKS;FLASHATTENTION_DISABLE_ALIBI -- CMAKE_PREFIX_PATH : /usr/lib/python3.13/site-packages -- CMAKE_INSTALL_PREFIX : /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch -- USE_GOLD_LINKER : OFF -- -- TORCH_VERSION : 2.4.0 -- BUILD_STATIC_RUNTIME_BENCHMARK: OFF -- BUILD_BINARY : OFF -- BUILD_CUSTOM_PROTOBUF : OFF -- Protobuf compiler : /usr/bin/protoc -- Protobuf includes : /usr/include -- Protobuf libraries : /usr/lib64/libprotobuf.so -- BUILD_DOCS : OFF -- BUILD_PYTHON : True -- Python version : 3.13.1 -- Python executable : /usr/bin/python3 -- Python library : -- Python includes : /usr/include/python3.13 -- Python site-package : /usr/lib/python3.13/site-packages -- BUILD_SHARED_LIBS : ON -- CAFFE2_USE_MSVC_STATIC_RUNTIME : OFF -- BUILD_TEST : False -- BUILD_JNI : OFF -- BUILD_MOBILE_AUTOGRAD : OFF -- BUILD_LITE_INTERPRETER: OFF -- INTERN_BUILD_MOBILE : -- TRACING_BASED : OFF -- USE_BLAS : 1 -- BLAS : flexi -- BLAS_HAS_SBGEMM : -- USE_LAPACK : 1 -- LAPACK : flexi -- USE_ASAN : OFF -- USE_TSAN : OFF -- USE_CPP_CODE_COVERAGE : OFF -- USE_CUDA : OFF -- USE_XPU : OFF -- USE_ROCM : ON -- ROCM_VERSION : -- USE_FLASH_ATTENTION : OFF -- USE_MEM_EFF_ATTENTION : OFF -- BUILD_NVFUSER : OFF -- USE_EIGEN_FOR_BLAS : ON -- USE_FBGEMM : OFF -- USE_FAKELOWP : OFF -- USE_KINETO : OFF -- USE_GFLAGS : OFF -- USE_GLOG : OFF -- USE_LITE_PROTO : OFF -- USE_PYTORCH_METAL : OFF -- USE_PYTORCH_METAL_EXPORT : OFF -- USE_MPS : OFF -- USE_MKL : OFF -- USE_MKLDNN : OFF -- USE_UCC : OFF -- USE_ITT : OFF -- USE_NCCL : OFF -- USE_NNPACK : OFF -- USE_NUMPY : ON -- USE_OBSERVERS : ON -- USE_OPENCL : OFF -- USE_OPENMP : ON -- USE_MIMALLOC : OFF -- USE_VULKAN : OFF -- USE_PROF : OFF -- USE_PYTORCH_QNNPACK : OFF -- USE_XNNPACK : ON -- USE_DISTRIBUTED : ON -- USE_MPI : OFF -- USE_GLOO : OFF -- USE_GLOO_WITH_OPENSSL : OFF -- USE_TENSORPIPE : ON -- Public Dependencies : -- Private Dependencies : Threads::Threads;pthreadpool;cpuinfo;XNNPACK;fp16;caffe2::openmp;tensorpipe;onnx_proto;onnx;rt;fmt;gcc_s;gcc;dl -- Public CUDA Deps. : -- Private CUDA Deps. : -- USE_COREML_DELEGATE : OFF -- BUILD_LAZY_TS_BACKEND : ON -- USE_ROCM_KERNEL_ASSERT : OFF -- Performing Test HAS_WMISSING_PROTOTYPES -- Performing Test HAS_WMISSING_PROTOTYPES - Failed -- Performing Test HAS_WERROR_MISSING_PROTOTYPES -- Performing Test HAS_WERROR_MISSING_PROTOTYPES - Failed -- Configuring done (34.3s) -- Generating done (0.6s) -- Build files have been written to: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build [1/4] Generating ATen headers [2/4] Generating ATen sources [3/4] Generating ATen declarations_yaml [1/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/fs-poll.c.o [2/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/idna.c.o [3/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/random.c.o [4/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/strscpy.c.o [5/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/inet.c.o [6/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/threadpool.c.o [7/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/timer.c.o [8/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/uv-data-getter-setters.c.o [9/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/version.c.o [10/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/async.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/internal.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c: In function ‘uv__async_io’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_22(D) + 432B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:149:3: note: in expansion of macro ‘QUEUE_MOVE’ 149 | QUEUE_MOVE(&loop->async_handles, &queue); | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:125:9: note: ‘queue’ declared here 125 | QUEUE queue; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:122:37: note: ‘loop’ declared here 122 | static void uv__async_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) { | ~~~~~~~~~~~^~~~ [11/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/uv-common.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c: In function ‘uv_walk’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_13(D) + 16B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:491:3: note: in expansion of macro ‘QUEUE_MOVE’ 491 | QUEUE_MOVE(&loop->handle_queue, &queue); | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:487:9: note: ‘queue’ declared here 487 | QUEUE queue; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:486:25: note: ‘loop’ declared here 486 | void uv_walk(uv_loop_t* loop, uv_walk_cb walk_cb, void* arg) { | ~~~~~~~~~~~^~~~ [12/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/dl.c.o [13/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/core.c.o [14/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/getaddrinfo.c.o [15/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/getnameinfo.c.o [16/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/loop-watcher.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/internal.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c: In function ‘uv__run_prepare’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_12(D) + 384B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:52:5: note: in expansion of macro ‘QUEUE_MOVE’ 52 | QUEUE_MOVE(&loop->name##_handles, &queue); \ | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:66:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 66 | UV_LOOP_WATCHER_DEFINE(prepare, PREPARE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:50:11: note: ‘queue’ declared here 50 | QUEUE queue; \ | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:66:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 66 | UV_LOOP_WATCHER_DEFINE(prepare, PREPARE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:48:34: note: ‘loop’ declared here 48 | void uv__run_##name(uv_loop_t* loop) { \ | ~~~~~~~~~~~^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:66:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 66 | UV_LOOP_WATCHER_DEFINE(prepare, PREPARE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c: In function ‘uv__run_check’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_12(D) + 400B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:52:5: note: in expansion of macro ‘QUEUE_MOVE’ 52 | QUEUE_MOVE(&loop->name##_handles, &queue); \ | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:67:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 67 | UV_LOOP_WATCHER_DEFINE(check, CHECK) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:50:11: note: ‘queue’ declared here 50 | QUEUE queue; \ | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:67:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 67 | UV_LOOP_WATCHER_DEFINE(check, CHECK) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:48:34: note: ‘loop’ declared here 48 | void uv__run_##name(uv_loop_t* loop) { \ | ~~~~~~~~~~~^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:67:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 67 | UV_LOOP_WATCHER_DEFINE(check, CHECK) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c: In function ‘uv__run_idle’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_12(D) + 416B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:52:5: note: in expansion of macro ‘QUEUE_MOVE’ 52 | QUEUE_MOVE(&loop->name##_handles, &queue); \ | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:68:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 68 | UV_LOOP_WATCHER_DEFINE(idle, IDLE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:50:11: note: ‘queue’ declared here 50 | QUEUE queue; \ | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:68:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 68 | UV_LOOP_WATCHER_DEFINE(idle, IDLE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:48:34: note: ‘loop’ declared here 48 | void uv__run_##name(uv_loop_t* loop) { \ | ~~~~~~~~~~~^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:68:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 68 | UV_LOOP_WATCHER_DEFINE(idle, IDLE) | ^~~~~~~~~~~~~~~~~~~~~~ [17/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/fs.c.o [18/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/loop.c.o [19/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/pipe.c.o [20/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/poll.c.o [21/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/random-devurandom.c.o [22/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/process.c.o [23/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/signal.c.o [24/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/stream.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/internal.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c: In function ‘uv__write_callbacks’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘pq’ in ‘((void * (**)[2])MEM[(void *[2] * *)stream_23(D) + 208B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:935:3: note: in expansion of macro ‘QUEUE_MOVE’ 935 | QUEUE_MOVE(&stream->write_completed_queue, &pq); | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:930:9: note: ‘pq’ declared here 930 | QUEUE pq; | ^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:927:46: note: ‘stream’ declared here 927 | static void uv__write_callbacks(uv_stream_t* stream) { | ~~~~~~~~~~~~~^~~~~~ [25/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/tcp.c.o [26/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/thread.c.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c: In function ‘thread_stack_size’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c:195:24: warning: comparison of integer expressions of different signedness: ‘rlim_t’ {aka ‘long unsigned int’} and ‘long int’ [-Wsign-compare] 195 | if (lim.rlim_cur >= PTHREAD_STACK_MIN) | ^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c: In function ‘uv_thread_create_ex’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c:243:20: warning: comparison of integer expressions of different signedness: ‘size_t’ {aka ‘long unsigned int’} and ‘long int’ [-Wsign-compare] 243 | if (stack_size < PTHREAD_STACK_MIN) | ^ [27/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/tty.c.o [28/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/proctitle.c.o [29/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/udp.c.o [30/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/linux-core.c.o [31/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/linux-syscalls.c.o [32/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/procfs-exepath.c.o [33/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/linux-inotify.c.o [34/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/random-getrandom.c.o [35/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/random-sysctl-linux.c.o [36/1912] Building CXX object c10/CMakeFiles/c10.dir/core/AutogradState.cpp.o [37/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Allocator.cpp.o [38/1912] Building CXX object c10/CMakeFiles/c10.dir/core/ConstantSymNodeImpl.cpp.o [39/1912] Building CXX object c10/CMakeFiles/c10.dir/core/CPUAllocator.cpp.o [40/1912] Building CXX object c10/CMakeFiles/c10.dir/core/CopyBytes.cpp.o [41/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DefaultDtype.cpp.o [42/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DeviceType.cpp.o [43/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Device.cpp.o [44/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DispatchKeySet.cpp.o [45/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DispatchKey.cpp.o [46/1912] Building CXX object c10/CMakeFiles/c10.dir/core/GradMode.cpp.o [47/1912] Building CXX object c10/CMakeFiles/c10.dir/core/InferenceMode.cpp.o [48/1912] Building CXX object c10/CMakeFiles/c10.dir/core/GeneratorImpl.cpp.o [49/1912] Building CXX object c10/CMakeFiles/c10.dir/core/RefcountedDeleter.cpp.o [50/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SafePyObject.cpp.o [51/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Scalar.cpp.o [52/1912] Building CXX object c10/CMakeFiles/c10.dir/core/ScalarType.cpp.o [53/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Storage.cpp.o [54/1912] Building CXX object c10/CMakeFiles/c10.dir/core/StorageImpl.cpp.o [55/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Stream.cpp.o [56/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymBool.cpp.o [57/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymFloat.cpp.o [58/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymIntArrayRef.cpp.o [59/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymNodeImpl.cpp.o [60/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymInt.cpp.o [61/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymbolicShapeMeta.cpp.o [62/1912] Building CXX object c10/CMakeFiles/c10.dir/core/TensorOptions.cpp.o [63/1912] Building CXX object c10/CMakeFiles/c10.dir/core/UndefinedTensorImpl.cpp.o [64/1912] Building CXX object c10/CMakeFiles/c10.dir/core/WrapDimMinimal.cpp.o [65/1912] Building CXX object c10/CMakeFiles/c10.dir/core/TensorImpl.cpp.o [66/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/COW.cpp.o [67/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/COWDeleter.cpp.o [68/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/DeviceGuardImplInterface.cpp.o [69/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/HermeticPyObjectTLS.cpp.o [70/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/GPUTrace.cpp.o [71/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/LocalDispatchKeySet.cpp.o [72/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/PyInterpreter.cpp.o [73/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/PyObjectSlot.cpp.o [74/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/PythonDispatcherTLS.cpp.o [75/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/SizesAndStrides.cpp.o [76/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/TorchDispatchModeTLS.cpp.o [77/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/alloc_cpu.cpp.o [78/1912] Building CXX object c10/CMakeFiles/c10.dir/core/thread_pool.cpp.o [79/1912] Building CXX object c10/CMakeFiles/c10.dir/mobile/CPUCachingAllocator.cpp.o [80/1912] Building CXX object c10/CMakeFiles/c10.dir/util/ApproximateClock.cpp.o [81/1912] Building CXX object c10/CMakeFiles/c10.dir/mobile/CPUProfilingAllocator.cpp.o [82/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Bfloat16.cpp.o [83/1912] Building CXX object c10/CMakeFiles/c10.dir/util/C++17.cpp.o [84/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Backtrace.cpp.o [85/1912] Building CXX object c10/CMakeFiles/c10.dir/util/DeadlockDetection.cpp.o [86/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e4m3fn.cpp.o [87/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e4m3fnuz.cpp.o [88/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e5m2.cpp.o [89/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Exception.cpp.o [90/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e5m2fnuz.cpp.o [91/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Half.cpp.o [92/1912] Building CXX object c10/CMakeFiles/c10.dir/util/LeftRight.cpp.o [93/1912] Building CXX object c10/CMakeFiles/c10.dir/util/MathConstants.cpp.o [94/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Metaprogramming.cpp.o [95/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Optional.cpp.o [96/1912] Building CXX object c10/CMakeFiles/c10.dir/util/ParallelGuard.cpp.o [97/1912] Building CXX object c10/CMakeFiles/c10.dir/util/SmallVector.cpp.o [98/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Logging.cpp.o [99/1912] Building CXX object c10/CMakeFiles/c10.dir/util/StringUtil.cpp.o [100/1912] Building CXX object c10/CMakeFiles/c10.dir/util/ThreadLocalDebugInfo.cpp.o [101/1912] Building CXX object c10/CMakeFiles/c10.dir/util/TypeCast.cpp.o [102/1912] Building CXX object c10/CMakeFiles/c10.dir/util/TypeTraits.cpp.o [103/1912] Building CXX object c10/CMakeFiles/c10.dir/util/TypeList.cpp.o [104/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Type_no_demangle.cpp.o [105/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Unicode.cpp.o [106/1912] Building CXX object c10/CMakeFiles/c10.dir/util/UniqueVoidPtr.cpp.o [107/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Type_demangle.cpp.o [108/1912] Building CXX object c10/CMakeFiles/c10.dir/util/flags_use_gflags.cpp.o [109/1912] Building CXX object c10/CMakeFiles/c10.dir/util/complex_math.cpp.o [110/1912] Building CXX object c10/CMakeFiles/c10.dir/util/int128.cpp.o [111/1912] Building CXX object c10/CMakeFiles/c10.dir/util/flags_use_no_gflags.cpp.o [112/1912] Building CXX object c10/CMakeFiles/c10.dir/util/intrusive_ptr.cpp.o [113/1912] Building CXX object c10/CMakeFiles/c10.dir/util/numa.cpp.o [114/1912] Building CXX object c10/CMakeFiles/c10.dir/util/signal_handler.cpp.o [115/1912] Building CXX object c10/CMakeFiles/c10.dir/util/thread_name.cpp.o [116/1912] Building CXX object c10/CMakeFiles/c10.dir/util/tempfile.cpp.o [117/1912] Linking C static library lib/libtensorpipe_uv.a [118/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/error.cc.o [119/1912] Building CXX object c10/CMakeFiles/c10.dir/util/typeid.cpp.o [120/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/address.cc.o [121/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/helpers.cc.o [122/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/error.cc.o [123/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/allocator.cc.o [124/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/fd.cc.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc: In member function ‘tensorpipe::Error tensorpipe::Fd::readFull(void*, size_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc:50:10: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 50 | if (rv != count) { | ~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc: In member function ‘tensorpipe::Error tensorpipe::Fd::writeFull(const void*, size_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc:62:10: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 62 | if (rv != count) { | ~~~^~~~~~~~ [125/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/socket.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [126/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/system.cc.o [127/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/context.cc.o [128/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/error.cc.o [129/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/listener.cc.o [130/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/context_impl.cc.o [131/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/pipe.cc.o [132/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/listener_impl.cc.o [133/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/error.cc.o [134/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/basic/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::SendOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.cc:51:28: required from here 51 | sendOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::RecvOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.cc:118:28: required from here 118 | recvOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [135/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/basic/context_impl.cc.o [136/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/basic/factory.cc.o [137/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/pipe_impl.cc.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc: In function ‘void tensorpipe::{anonymous}::parseDescriptorReplyOfMessage(tensorpipe::WriteOperation&, tensorpipe::DescriptorReply)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:35:40: warning: comparison of integer expressions of different signedness: ‘size_t’ {aka ‘long unsigned int’} and ‘const int’ [-Wsign-compare] 35 | for (size_t tensorIdx = 0; tensorIdx < numTensors; ++tensorIdx) { | ~~~~~~~~~~^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc: In function ‘std::shared_ptr > tensorpipe::{anonymous}::makeDescriptorForMessage(const tensorpipe::WriteOperation&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:77:39: warning: comparison of integer expressions of different signedness: ‘int’ and ‘std::vector::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 77 | for (int payloadIdx = 0; payloadIdx < op.message.payloads.size(); | ~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:87:37: warning: comparison of integer expressions of different signedness: ‘int’ and ‘std::vector::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 87 | for (int tensorIdx = 0; tensorIdx < op.tensors.size(); ++tensorIdx) { | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc: In member function ‘void tensorpipe::PipeImpl::expectReadCall(ReadOpIter)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:869:18: warning: unused variable ‘op’ [-Wunused-variable] 869 | ReadOperation& op = *opIter; | ^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::ReadOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::ReadOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:391:28: required from here 391 | readOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.h:21: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::WriteOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::WriteOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:580:29: required from here 580 | writeOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/serializer.h:24, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/nop_types.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.h:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::pair; T = std::__cxx11::basic_string; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::__cxx11::basic_string > >; Type = std::unordered_map, std::__cxx11::basic_string >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::__cxx11::basic_string >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transport, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::address, void*, 0, void>, nop::MemberPointer, std::equal_to, std::allocator > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportDomainDescriptor, void*, 0, void>, nop::MemberPointer, std::allocator >, std::vector >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::vector > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelDeviceDescriptors, void*, 0, void>, nop::MemberPointer, std::__cxx11::basic_string, std::allocator >, std::hash >, std::equal_to >, std::allocator, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelForDevicePair, void*, 0, void> >; T = std::unordered_map, std::__cxx11::basic_string >; Class = tensorpipe::BrochureAnswer; T Class::* Pointer = &tensorpipe::BrochureAnswer::channelForDevicePair]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 7; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::BrochureAnswer]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::__cxx11::basic_string >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, false>::value_type’ {aka ‘const std::pair, std::__cxx11::basic_string >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::__cxx11::basic_string >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, false>::value_type&’ {aka ‘const std::pair, std::__cxx11::basic_string >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::unordered_map >; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::unordered_map > > >; Type = std::unordered_map, std::unordered_map > >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::unordered_map > >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::transportDomainDescriptors, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::channelDeviceDescriptors, void*, 0, void> >; T = std::unordered_map, std::unordered_map > >; Class = tensorpipe::Brochure; T Class::* Pointer = &tensorpipe::Brochure::channelDeviceDescriptors]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 2; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::Brochure]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::unordered_map > >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::unordered_map > >, false, true>::value_type’ {aka ‘const std::pair, std::unordered_map > >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::unordered_map > >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::unordered_map > >, false, true>::value_type&’ {aka ‘const std::pair, std::unordered_map > >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::__cxx11::basic_string; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::__cxx11::basic_string > >; Type = std::unordered_map, std::__cxx11::basic_string >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::__cxx11::basic_string >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::transportDomainDescriptors, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::channelDeviceDescriptors, void*, 0, void> >; T = std::unordered_map, std::__cxx11::basic_string >; Class = tensorpipe::Brochure; T Class::* Pointer = &tensorpipe::Brochure::transportDomainDescriptors]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 1; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 2; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 99 | auto status = WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::Brochure]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::__cxx11::basic_string >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, true>::value_type’ {aka ‘const std::pair, std::__cxx11::basic_string >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::__cxx11::basic_string >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, true>::value_type&’ {aka ‘const std::pair, std::__cxx11::basic_string >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::vector; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::vector > >; Type = std::unordered_map, std::vector >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::vector >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transport, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::address, void*, 0, void>, nop::MemberPointer, std::equal_to, std::allocator > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportDomainDescriptor, void*, 0, void>, nop::MemberPointer, std::allocator >, std::vector >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::vector > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelDeviceDescriptors, void*, 0, void>, nop::MemberPointer, std::__cxx11::basic_string, std::allocator >, std::hash >, std::equal_to >, std::allocator, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelForDevicePair, void*, 0, void> >; T = std::unordered_map, std::vector >; Class = tensorpipe::BrochureAnswer; T Class::* Pointer = &tensorpipe::BrochureAnswer::channelRegistrationIds]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: recursively required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 6; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 99 | auto status = WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 7; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::BrochureAnswer]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::vector >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::vector >, false, true>::value_type’ {aka ‘const std::pair, std::vector >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::vector >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::vector >, false, true>::value_type&’ {aka ‘const std::pair, std::vector >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = tensorpipe::Device; T = std::__cxx11::basic_string; Hash = std::hash; KeyEqual = std::equal_to; Allocator = std::allocator > >; Type = std::unordered_map >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:147:34: required from ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::unordered_map >; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::unordered_map > > >; Type = std::unordered_map, std::unordered_map > >]’ 147 | status = Encoding::Write(element.second, writer); | ~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::unordered_map > >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::transportDomainDescriptors, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::channelDeviceDescriptors, void*, 0, void> >; T = std::unordered_map, std::unordered_map > >; Class = tensorpipe::Brochure; T Class::* Pointer = &tensorpipe::Brochure::channelDeviceDescriptors]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 2; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::Brochure]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator >, false, false>::value_type’ {aka ‘const std::pair >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator >, false, false>::value_type&’ {aka ‘const std::pair >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = long unsigned int; T = long unsigned int; Hash = std::hash; KeyEqual = std::equal_to; Allocator = std::allocator >; Type = std::unordered_map]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transport, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::address, void*, 0, void>, nop::MemberPointer, std::equal_to, std::allocator > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportDomainDescriptor, void*, 0, void>, nop::MemberPointer, std::allocator >, std::vector >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::vector > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelDeviceDescriptors, void*, 0, void>, nop::MemberPointer, std::__cxx11::basic_string, std::allocator >, std::hash >, std::equal_to >, std::allocator, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelForDevicePair, void*, 0, void> >; T = std::unordered_map; Class = tensorpipe::BrochureAnswer; T Class::* Pointer = &tensorpipe::BrochureAnswer::transportRegistrationIds]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: recursively required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 6; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 99 | auto status = WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 7; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::BrochureAnswer]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, false, false>::value_type’ {aka ‘const std::pair’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, false, false>::value_type&’ {aka ‘const std::pair&’} to prevent copying [138/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/xth/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::SendOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.cc:65:28: required from here 65 | sendOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::RecvOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.cc:157:28: required from here 157 | recvOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [139/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/xth/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/context_impl.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/context_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h: In instantiation of ‘void tensorpipe::Queue::push(T) [with T = tensorpipe::optional]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/context_impl.cc:68:17: required from here 68 | requests_.push(nullopt); | ~~~~~~~~~~~~~~^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h:24:26: warning: comparison of integer expressions of different signedness: ‘std::deque, std::allocator > >::size_type’ {aka ‘long unsigned int’} and ‘const int’ [-Wsign-compare] 24 | while (items_.size() >= capacity_) { | ~~~~~~~~~~~~~~^~~~~~~~~~~~ [140/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/xth/factory.cc.o [141/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/cma/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::SendOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.cc:66:28: required from here 66 | sendOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::RecvOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.cc:160:28: required from here 160 | recvOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [142/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/cma/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/strings.h: In function ‘bool tensorpipe::isValidUuid(const std::string&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/strings.h:58:21: warning: comparison of integer expressions of different signedness: ‘int’ and ‘std::__cxx11::basic_string::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 58 | for (int i = 0; i < uuid.size(); i++) { | ~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc: In function ‘tensorpipe::Error tensorpipe::channel::cma::{anonymous}::callProcessVmReadv(void*, void*, size_t, pid_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:61:20: warning: comparison of integer expressions of different signedness: ‘ssize_t’ {aka ‘long int’} and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 61 | } else if (nread != length) { | ~~~~~~^~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h: In instantiation of ‘void tensorpipe::Queue::push(T) [with T = tensorpipe::optional]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:324:17: required from here 324 | requests_.push(nullopt); | ~~~~~~~~~~~~~~^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h:24:26: warning: comparison of integer expressions of different signedness: ‘std::deque, std::allocator > >::size_type’ {aka ‘long unsigned int’} and ‘const int’ [-Wsign-compare] 24 | while (items_.size() >= capacity_) { | ~~~~~~~~~~~~~~^~~~~~~~~~~~ [143/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/cma/factory.cc.o [144/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/mpt/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:74:20: required from ‘void tensorpipe::OpsStateMachine::advanceAllOperations() [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::SendOperation]’ 74 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.cc:131:32: required from here 131 | sendOps_.advanceAllOperations(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.h:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:74:20: required from ‘void tensorpipe::OpsStateMachine::advanceAllOperations() [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::RecvOperation]’ 74 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.cc:132:32: required from here 132 | recvOps_.advanceAllOperations(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [145/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/mpt/context_impl.cc.o [146/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/mpt/factory.cc.o [147/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/uv.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/loop.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/context_impl.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/context_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [148/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/error.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/uv.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/error.cc:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [149/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/connection_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/connection_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/connection_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [150/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/listener_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/listener_impl.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/listener_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [151/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/factory.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/connection_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/factory.cc:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [152/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/sockaddr.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [153/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/loop.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/uv.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/loop.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/loop.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [154/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/epoll_loop.cc.o [155/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/utility.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/utility.cc:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [156/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/shm_segment.cc.o [157/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/context_impl.cc:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [158/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/connection_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘tensorpipe::Error tensorpipe::recvFromSocket(int, T&, T&, Fds& ...) [with T = unsigned int; Fds = {Fd, Fd, Fd, Fd}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:253:26: required from ‘tensorpipe::Error tensorpipe::Socket::recvPayloadAndFds(T&, T&, Fds& ...) [with T = unsigned int; Fds = {tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd}; typename std::enable_if::value, bool>::type = false]’ 253 | return recvFromSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:219:41: required from here 219 | auto err = socket_.recvPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 220 | peerInboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~ 221 | peerOutboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~~ 222 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 223 | reactorDataFd, | ~~~~~~~~~~~~~~ 224 | outboxHeaderFd, | ~~~~~~~~~~~~~~~ 225 | outboxDataFd); | ~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:154:12: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 154 | if (rv != iov.iov_len) { | ~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘tensorpipe::Error tensorpipe::sendToSocket(int, const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:243:24: required from ‘tensorpipe::Error tensorpipe::Socket::sendPayloadAndFds(const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}; typename std::enable_if::value, bool>::type = false]’ 243 | return sendToSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:274:41: required from here 274 | auto err = socket_.sendPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 275 | inboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ 276 | outboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 277 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 278 | reactorDataFd, | ~~~~~~~~~~~~~~ 279 | inboxHeaderSegment_.getFd(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 280 | inboxDataSegment_.getFd()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:107:12: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 107 | if (rv != iov.iov_len) { | ~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘void tensorpipe::{anonymous}::loadFdsFromArray(int*, std::index_sequence, Fds& ...) [with long unsigned int ...Idxs = {0, 1, 2, 3}; Fds = {tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd}; std::index_sequence = std::integer_sequence]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:171:19: required from ‘tensorpipe::Error tensorpipe::recvFromSocket(int, T&, T&, Fds& ...) [with T = unsigned int; Fds = {Fd, Fd, Fd, Fd}]’ 171 | loadFdsFromArray(payload, std::index_sequence_for{}, fds...); | ~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:253:26: required from ‘tensorpipe::Error tensorpipe::Socket::recvPayloadAndFds(T&, T&, Fds& ...) [with T = unsigned int; Fds = {tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd}; typename std::enable_if::value, bool>::type = false]’ 253 | return recvFromSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:219:41: required from here 219 | auto err = socket_.recvPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 220 | peerInboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~ 221 | peerOutboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~~ 222 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 223 | reactorDataFd, | ~~~~~~~~~~~~~~ 224 | outboxHeaderFd, | ~~~~~~~~~~~~~~~ 225 | outboxDataFd); | ~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:58:8: warning: unused variable ‘dummy’ [-Wunused-variable] 58 | auto dummy = {(loadOneFdFromArray(array[Idxs], fds), 0)...}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘void tensorpipe::{anonymous}::saveFdsToArray(int*, std::index_sequence, const Fds& ...) [with long unsigned int ...Idxs = {0, 1, 2, 3}; Fds = {int, int, int, int}; std::index_sequence = std::integer_sequence]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:96:17: required from ‘tensorpipe::Error tensorpipe::sendToSocket(int, const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}]’ 96 | saveFdsToArray(payload, std::index_sequence_for{}, fds...); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:243:24: required from ‘tensorpipe::Error tensorpipe::Socket::sendPayloadAndFds(const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}; typename std::enable_if::value, bool>::type = false]’ 243 | return sendToSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:274:41: required from here 274 | auto err = socket_.sendPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 275 | inboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ 276 | outboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 277 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 278 | reactorDataFd, | ~~~~~~~~~~~~~~ 279 | inboxHeaderSegment_.getFd(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 280 | inboxDataSegment_.getFd()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:41:8: warning: unused variable ‘dummy’ [-Wunused-variable] 41 | auto dummy = {(saveOneFdToArray(array[Idxs], fds), 0)...}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ [159/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/factory.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/factory.cc:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [160/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/listener_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/listener_impl.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/listener_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [161/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/sockaddr.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/sockaddr.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/sockaddr.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [162/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/ibv.cc.o [163/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/reactor.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/busy_polling_loop.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h: In instantiation of ‘ssize_t tensorpipe::RingBufferRole::write(const void*, size_t) [with int NumRoles = 2; int RoleIdx = 1; ssize_t = long int; size_t = long unsigned int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.cc:22:29: required from here 22 | auto rv = producer.write(&token, sizeof(token)); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h:270:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 270 | TP_DCHECK_EQ(ret, size); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h: In instantiation of ‘ssize_t tensorpipe::RingBufferRole::read(void*, size_t) [with int NumRoles = 2; int RoleIdx = 0; ssize_t = long int; size_t = long unsigned int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.cc:114:34: required from here 114 | auto ret = reactorConsumer.read(&token, sizeof(token)); | ~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘const size_t’ {aka ‘const long unsigned int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h:248:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 248 | TP_DCHECK_EQ(ret, size); | ^~~~~~~~~~~~ [164/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/context_impl.cc:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [165/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/connection_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [166/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/error.cc.o [167/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/listener_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/listener_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/listener_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [168/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/factory.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/factory.cc:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [169/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/sockaddr.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/sockaddr.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [170/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/utility.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/utility.cc:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [171/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/reactor.cc.o [172/1912] Linking CXX shared library lib/libc10.so [173/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx.dir/common_avx.cc.o [174/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx.dir/typed_axpy_avx.cc.o [175/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/adagrad_avx2.cc.o [176/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/batch_box_cox_avx2.cc.o [177/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/common_avx2.cc.o [178/1912] Generating /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/Functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_3.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_4.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_3.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_4.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ADInplaceOrViewType_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ADInplaceOrViewType_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNativeFunctions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterAutogradLazy.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterLazy.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/Functions.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/variable_factories.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyIr.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNonNativeIr.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNativeFunctions.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_3.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_4.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_variable_methods.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_nn_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_fft_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_linalg_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_nested_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_sparse_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_special_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_return_types.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_enum_tag.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_return_types.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/testing/_internal/generated/annotated_fn_args.py, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.cpp [179/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_avx2.cc.o [180/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_fused_8bit_rowwise_avx2.cc.o [181/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_fused_8bit_rowwise_idx_avx2.cc.o [182/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/lstm_unit_cpu_avx2.cc.o [183/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/math_cpu_avx2.cc.o [184/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_idx_avx2.cc.o [185/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/typed_axpy_avx2.cc.o [186/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx512.dir/common_avx512.cc.o [187/1912] Linking C static library lib/libfp16.a [188/1912] Linking CXX static library lib/libtensorpipe.a [189/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx512.dir/adagrad_avx512.cc.o [190/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPAllocatorConfig.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [191/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPDeviceAssertionHost.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [192/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPException.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [193/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPFunctions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [194/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPMallocAsyncAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [195/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPMiscFunctions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [196/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPStream.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [197/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/driver_api.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [198/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/impl/HIPGuardImpl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [199/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/impl/HIPTest.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [200/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/AccumulateType.cpp.o [201/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPCachingAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [202/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/CPUGeneratorImpl.cpp.o [203/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ConjugateFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MathBitsFallback.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ConjugateFallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [204/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/CachedTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CachedTensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [205/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Context.cpp.o [206/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/DeviceAccelerator.cpp.o [207/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Dispatch.cpp.o [208/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/DynamicLibrary.cpp.o [209/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/EmptyTensor.cpp.o [210/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/DLConvertor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DLConvertor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DLConvertor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [211/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FuncTorchTLS.cpp.o [212/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ExpandUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [213/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalStorageImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalStorageImpl.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [214/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalInverses.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalInverses.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [215/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalTensorWrapper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [216/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalizeFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalizeFallbackKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [217/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyBatchedTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedTensorImpl.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [218/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyBatchedFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedFallback.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedFallback.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [219/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyVmapMode.cpp.o [220/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyVmapTransforms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyVmapTransforms.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyVmapTransforms.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [221/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/MapAllocator.cpp.o [222/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/MemoryOverlap.cpp.o [223/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/NamedTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [224/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyBatchingRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchingRegistrations.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [225/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelCommon.cpp.o [226/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelNative.cpp.o [227/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelOpenMP.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelFuture.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelOpenMP.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [228/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/NestedTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NestedTensorImpl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [229/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/PythonTorchFunctionTLS.cpp.o [230/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SavedTensorHooks.cpp.o [231/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelThreadPoolNative.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelThreadPoolNative.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [232/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SequenceNumber.cpp.o [233/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ScalarOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [234/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SparseCsrTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorImpl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [235/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SparseTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [236/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorGeometry.cpp.o [237/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/StorageUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/StorageUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [238/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorIndexing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [239/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorMeta.cpp.o [240/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorNames.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [241/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorIterator.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIterator.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [242/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ThreadLocalPythonObjects.cpp.o [243/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ThreadLocalState.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [244/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [245/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Version.cpp.o [246/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Utils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [247/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/VmapModeRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/VmapModeRegistrations.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [248/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ZeroTensorFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp: In function ‘void at::zeroTensorFallback(const c10::OperatorHandle&, c10::DispatchKeySet, torch::jit::Stack*)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:69:25: warning: possibly dangling reference to a temporary [-Wdangling-reference] 69 | const Tensor& tensor = tensors[j]; | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:69:43: note: the temporary was destroyed at the end of the full expression ‘tensors.c10::List::operator[](((long unsigned int)j)).c10::impl::ListElementReference > >::operator std::conditional_t()’ 69 | const Tensor& tensor = tensors[j]; | ^ [249/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/cpu/FlushDenormal.cpp.o [250/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/cpu/Utils.cpp.o [251/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/CPUGuardImpl.cpp.o [252/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/CUDAHooksInterface.cpp.o [253/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/HIPHooksInterface.cpp.o [254/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/IPUHooksInterface.cpp.o [255/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MAIAHooksInterface.cpp.o [256/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MPSHooksInterface.cpp.o [257/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MTIAHooksInterface.cpp.o [258/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MetaGuardImpl.cpp.o [259/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/PrivateUse1HooksInterface.cpp.o [260/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/XPUHooksInterface.cpp.o [261/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/ADInterpreters.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/ADInterpreters.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/ADInterpreters.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [262/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/autocast_mode.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [263/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesActivation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesActivation.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [264/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesBinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesBinaryOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [265/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesConvolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesConvolution.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [266/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesDynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesDynamic.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesDynamic.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [267/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesDecompositions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesDecompositions.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [268/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesFactory.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesFactory.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [269/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesHelper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [270/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesLoss.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesLoss.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [271/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesLinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesLinearAlgebra.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [272/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesModules.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesModules.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [273/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesNorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesNorm.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [274/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesPooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesPooling.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [275/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesRandomness.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesRandomness.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesRandomness.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [276/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesReduceOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesReduceOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [277/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesScatterOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesScatterOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [278/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesUnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesUnaryOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [279/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesViews.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesViews.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [280/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchedTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedTensorImpl.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [281/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchedFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedFallback.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedFallback.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedFallback.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [282/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/FunctionalizeInterpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/FunctionalizeInterpreter.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/FunctionalizeInterpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [283/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/Interpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [284/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/DynamicLayer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [285/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/LegacyVmapTransforms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/LegacyVmapTransforms.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [286/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/LegacyBatchingRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/LegacyBatchingRegistrations.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [287/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/PlumbingHelper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/TensorWrapper.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/PlumbingHelper.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [288/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/TensorWrapper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/TensorWrapper.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/TensorWrapper.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [289/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/PyTorchOperatorHacks.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/PyTorchOperatorHacks.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [290/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/VmapInterpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/VmapInterpreter.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/VmapInterpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [291/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/record_function.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [292/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/ATenGeneral.cpp.o [293/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/BackendSelectFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/BackendSelectFallbackKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [294/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/VmapModeRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/VmapModeRegistrations.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [295/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/DeprecatedTypePropertiesRegistry.cpp.o [296/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/DeprecatedTypeProperties.cpp.o [297/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Dimname.cpp.o [298/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Dict.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [299/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Formatting.cpp.o [300/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Generator.cpp.o [301/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/GeneratorForPrivateuseone.cpp.o [302/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/List.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [303/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/MetaFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/MetaFallbackKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [304/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/NamedTensor.cpp.o [305/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/NestedIntSymNodeImpl.cpp.o [306/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/PythonFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonFallbackKernel.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonFallbackKernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [307/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/NamedRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/NamedRegistrations.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [308/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Range.cpp.o [309/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/PythonOpRegistrationTrampoline.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonOpRegistrationTrampoline.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonOpRegistrationTrampoline.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [310/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Tensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [311/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/TorchDispatchUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [312/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/VariableHooksInterface.cpp.o [313/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Vitals.cpp.o [314/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/VariableFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/VariableFallbackKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [315/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/blob.cpp.o [316/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/adaption.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/adaption.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/adaption.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [317/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/boxing/KernelFunction.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [318/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/class_type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [319/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/custom_class.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/custom_class.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/custom_class.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [320/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/DispatchKeyExtractor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [321/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/ObservedOperators.cpp.o [322/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/Dispatcher.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [323/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/OperatorEntry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [324/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dynamic_type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dynamic_type.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [325/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/function_schema.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [326/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/interned_strings.cpp.o [327/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/library.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/library.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [328/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/ivalue.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [329/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/op_registration/infer_schema.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [330/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/operator_name.cpp.o [331/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/register_symbols.cpp.o [332/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/op_registration/op_registration.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/op_registration.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [333/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/tensor_type.cpp.o [334/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/type_factory.cpp.o [335/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [336/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/error_report.cpp.o [337/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/union_type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/union_type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [338/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/lexer.cpp.o [339/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/function_schema_parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [340/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/strtod.cpp.o [341/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/schema_type_parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_type_parser.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [342/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/source_range.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_range.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [343/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Activation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Activation.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Activation.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [344/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveAveragePooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_adaptive_avg_pool2d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveAveragePooling.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [345/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveMaxPooling2d.cpp.o [346/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveAveragePooling3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_adaptive_avg_pool3d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveAveragePooling3d.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveAveragePooling3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [347/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveMaxPooling3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveMaxPooling3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [348/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AffineGridGenerator.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AffineGridGenerator.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [349/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AmpKernels.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_amp_foreach_non_finite_check_and_unscale.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AmpKernels.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [350/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AutogradComposite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/alias.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AutogradComposite.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [351/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AveragePool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AveragePool2d.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [352/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AveragePool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AveragePool3d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [353/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BatchLinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BatchLinearAlgebra.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [354/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BatchLinearAlgebraKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BatchLinearAlgebraKernel.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [355/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BinaryOps.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [356/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Blas.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Blas.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Blas.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [357/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BlasKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BlasKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [358/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/CPUBlas.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Matmul.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUBlas.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/OpMathType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUBlas.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUBlas.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [359/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Bucketization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BucketizationUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Bucketization.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Bucketization.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [360/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ChanelShuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ChanelShuffle.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [361/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/CPUFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [362/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ComparisonUtils.cpp.o [363/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Col2Im.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Col2Im.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Col2Im.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [364/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Constraints.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Constraints.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [365/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ConvolutionMM2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_slow_conv2d_forward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionMM2d.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionMM2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [366/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Convolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Convolution.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [367/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ConvolutionTBC.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionTBC.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [368/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ConvolutionMM3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionMM3d.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [369/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Copy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Copy.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Copy.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [370/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Correlation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Correlation.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [371/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Cross.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Cross.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [372/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/DilatedMaxPool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/DilatedMaxPool2d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [373/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/DispatchStub.cpp.o [374/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/DilatedMaxPool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/DilatedMaxPool3d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [375/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Distance.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Distance.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [376/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Dropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Dropout.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [377/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Distributions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Distributions.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Distributions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [378/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Embedding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Embedding.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [379/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Fill.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Fill.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [380/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/EmbeddingBag.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/EmbeddingBag.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/EmbeddingBag.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [381/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FractionalMaxPool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FractionalMaxPool2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [382/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ForeachOpsKernels.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/copy.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ForeachOpsKernels.cpp:57: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ForeachOpsKernels.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [383/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FunctionOfAMatrixUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FunctionOfAMatrixUtils.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [384/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FractionalMaxPool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FractionalMaxPool3d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FractionalMaxPool3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [385/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FusedAdagrad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fused_adagrad.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FusedAdagrad.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [386/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FusedAdam.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fused_adam.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FusedAdam.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [387/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FusedSGD.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fused_sgd.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FusedSGD.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [388/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/GatedLinearUnit.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/GatedLinearUnit.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [389/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/GridSampler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/GridSampler.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [390/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Histogram.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Histogram.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [391/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/IndexingUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/IndexingUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/IndexingUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [392/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Im2Col.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Im2Col.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [393/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Integration.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Integration.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [394/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Itertools.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Itertools.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [395/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LegacyBridge.cpp.o [396/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Lerp.cpp.o [397/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LegacyBatching.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LegacyBatching.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [398/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Linear.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [399/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Loss.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Loss.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Loss.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [400/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebra.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebra.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [401/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossMultiLabelMargin.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossMultiLabelMargin.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [402/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossCTC.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossCTC.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [403/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossMultiMargin.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossMultiMargin.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [404/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossNLL.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [405/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossNLL2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL2d.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [406/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/MaxPooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MaxPooling.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [407/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/MaxUnpooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MaxUnpooling.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [408/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Memory.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_pin_memory.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Memory.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [409/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/MetaTensor.cpp.o [410/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NNPACK.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NNPACK.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [411/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NaiveConvolutionTranspose2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose2d.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [412/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NaiveConvolutionTranspose3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose3d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [413/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NamedTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NamedTensor.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [414/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NaiveDilatedConvolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveDilatedConvolution.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveDilatedConvolution.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [415/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NegateFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MathBitsFallback.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NegateFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [416/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Onehot.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Onehot.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [417/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PackedSequence.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/PackedSequence.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [418/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Normalization.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Normalization.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [419/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PadNd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/PadNd.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [420/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PointwiseOps.cpp.o [421/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PixelShuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/roll.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorTransformations.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/PixelShuffle.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [422/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Pooling.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [423/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Pow.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Pow.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [424/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/QuantizedLinear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtilsMulti.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/QuantizedLinear.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [425/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/RangeFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/linspace.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [426/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReduceAllOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceAllOps.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [427/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/RNN.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RNN.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RNN.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [428/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReflectionPad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReflectionPad.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [429/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReduceOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOps.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [430/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Repeat.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Repeat.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Repeat.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [431/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReplicationPadding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReplicationPadding.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [432/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/RowwisePrune.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RowwisePrune.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [433/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Resize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [434/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Scalar.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_local_scalar_dense.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Scalar.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Scalar.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [435/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SobolEngineOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOpsUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [436/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SegmentReduce.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SegmentReduce.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReductionType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SegmentReduce.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SegmentReduce.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [437/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SobolEngineOpsUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOpsUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOpsUtils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [438/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SoftMax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SoftMax.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SoftMax.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [439/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Sorting.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Sorting.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Sorting.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [440/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SparseTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [441/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SummaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SummaryOps.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [442/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SpectralOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SpectralOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [443/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorCompare.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorCompare.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorCompare.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [444/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorAdvancedIndexing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.cpp:51: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [445/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorConversions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorConversions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [446/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorIteratorReduce.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorIteratorReduce.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [447/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [448/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorProperties.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorProperties.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [449/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorTransformations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/roll.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorTransformations.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorTransformations.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [450/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TestOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TestOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [451/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TriangularOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TriangularOpsUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TriangularOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [452/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorShape.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorShape.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [453/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TypeProperties.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TypeProperties.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [454/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Unfold2d.cpp.o [455/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnaryOps.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [456/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Unfold3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Unfold3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [457/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UnfoldBackward.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [458/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSample.cpp.o [459/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleBicubic2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_bicubic2d_aa.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleBicubic2d.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleBicubic2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [460/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Unique.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Unique.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Unique.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [461/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleLinear1d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/upsample_linear1d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleLinear1d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [462/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleBilinear2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_bilinear2d_aa.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleBilinear2d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [463/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleNearest1d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_nearest_exact1d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleNearest1d.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [464/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleNearest2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_nearest_exact2d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleNearest2d.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [465/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleNearest3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_nearest_exact3d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleNearest3d.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [466/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleTrilinear3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/upsample_trilinear3d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleTrilinear3d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [467/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/VariableMethodStubs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/alias.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/VariableMethodStubs.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [468/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/WeightNorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/WeightNorm.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [469/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/group_norm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/group_norm.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [470/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/layer_norm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [471/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/verbose_wrapper.cpp.o [472/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/prim_native_functions.cpp.o [473/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/library.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/library.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [474/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/fbgemm_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/fbgemm_utils.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [475/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [476/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_dynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_dynamic.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [477/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_deserialize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_deserialize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [478/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_prepack.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [479/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_serialize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_serialize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [480/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_unpack.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [481/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/ParamUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/ParamUtils.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [482/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/FlattenIndicesKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/FlattenIndicesCommon.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/FlattenIndicesKernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [483/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SoftMax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SoftMax.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [484/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseBlas.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBlas.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [485/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionCommon.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionKernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionCommon.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [486/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseBlasImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBlasImpl.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBlasImpl.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [487/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseCsrTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseCsrTensor.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [488/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_unique.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseFactories.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [489/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseMatMul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseMatMul.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [490/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseTensor.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseTensor.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [491/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseCsrTensorMath.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseCsrTensorMath.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseCsrTensorMath.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [492/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseUnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseUnaryOps.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [493/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseTensorMath.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseTensorMath.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [494/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/ValidateCompressedIndicesKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/ValidateCompressedIndicesCommon.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/ValidateCompressedIndicesKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [495/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorAliases.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorAliases.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [496/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorBackward.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorBackward.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [497/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorBinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorBinaryOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [498/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorFactories.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [499/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorMatmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUtils.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMatmul.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [500/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorMath.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMath.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NestedTensorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMath.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMath.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [501/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorUnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUnaryOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [502/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorTransformerFunctions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorTransformerFunctions.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [503/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUtils.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [504/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/AffineQuantizer.cpp.o [505/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/AffineQuantizerBase.cpp.o [506/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/Copy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/Copy.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [507/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/FakeQuantPerChannelAffine.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/FakeQuantPerChannelAffine.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [508/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/FakeQuantPerTensorAffine.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/FakeQuantPerTensorAffine.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [509/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/QTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/QTensor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [510/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/TensorAdvancedIndexing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/TensorAdvancedIndexing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [511/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/TensorCompare.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/TensorCompare.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [512/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/AdaptiveAveragePooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/AdaptiveAveragePooling.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [513/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/TensorFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/TensorFactories.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [514/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/AveragePool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/AveragePool2d.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [515/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/AveragePool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/AveragePool3d.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [516/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/ChannelShuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/ChannelShuffle.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [517/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/BinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/BinaryOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/BinaryOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [518/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/IntReprQuant.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/IntReprQuant.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [519/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/LinearUnpackImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpp_custom_type_hack.h:57, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/LinearUnpackImpl.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/LinearUnpackImpl.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [520/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/MakePerTensorQuantizedTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/MakePerTensorQuantizedTensor.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [521/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/Normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/Normalization.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [522/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/ReduceOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/ReduceOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [523/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/RuyUtils.cpp.o [524/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/Pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/Pooling.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [525/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/Sorting.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/Sorting.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [526/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/TensorOperators.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/TensorOperators.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [527/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/UpSampleBilinear2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/UpSampleBilinear2d.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [528/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/TensorShape.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/TensorShape.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/TensorShape.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [529/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/UpSampleNearest2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/UpSampleNearest2d.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [530/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/UpSampleNearest3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/UpSampleNearest3d.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [531/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/XnnpackUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/XnnpackUtils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [532/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:21: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h: In instantiation of ‘c10::intrusive_ptr > deserialize_conv(ConvParamsSerializationTypeV3) [with unsigned int kSpatialDim = 2; ConvParamsSerializationTypeV3 = std::tuple >, std::vector, std::allocator > > >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:410:47: required from ‘int register_conv_params() [with int kSpatialDim = 2]’ 410 | return deserialize_conv(state); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:429:39: required from here 429 | TORCH_API int register_conv_params<2>(); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:316:11: warning: unused variable ‘groups’ [-Wunused-variable] 316 | int64_t groups = config_vals.at(idx); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:326:8: warning: unused variable ‘transpose’ [-Wunused-variable] 326 | bool transpose = flags & (1 << 0); | ^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h: In instantiation of ‘c10::intrusive_ptr > deserialize_conv(ConvParamsSerializationTypeV3) [with unsigned int kSpatialDim = 3; ConvParamsSerializationTypeV3 = std::tuple >, std::vector, std::allocator > > >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:410:47: required from ‘int register_conv_params() [with int kSpatialDim = 3]’ 410 | return deserialize_conv(state); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:431:39: required from here 431 | TORCH_API int register_conv_params<3>(); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:316:11: warning: unused variable ‘groups’ [-Wunused-variable] 316 | int64_t groups = config_vals.at(idx); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:326:8: warning: unused variable ‘transpose’ [-Wunused-variable] 326 | bool transpose = flags & (1 << 0); | ^~~~~~~~~ [533/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/init_qnnpack.cpp.o [534/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/fused_obs_fake_quant.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fused_obs_fake_quant.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [535/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qclamp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qclamp.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [536/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [537/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv_dynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_dynamic.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_dynamic.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [538/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_prepack.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_prepack.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [539/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv_unpack_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_unpack_impl.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_unpack_impl.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [540/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qdropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qdropout.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qdropout.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [541/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qelu.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qelu.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [542/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qembeddingbag.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/EmbeddingPackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [543/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qembeddingbag_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/EmbeddingPackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_prepack.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_prepack.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [544/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qgelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qgelu.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [545/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qembeddingbag_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/EmbeddingPackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_unpack.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_unpack.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [546/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qhardsigmoid.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qhardsigmoid.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [547/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qhardswish.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qhardswish.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [548/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qlinear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [549/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qlinear_dynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/PackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_dynamic.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_dynamic.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [550/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qlinear_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/PackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_prepack.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_prepack.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [551/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmul.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmul.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [552/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qmatmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmatmul.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmatmul.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [553/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qnormalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qnormalization.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qnormalization.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [554/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qrelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qrelu.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qrelu.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [555/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qsigmoid.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qsigmoid.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [556/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qtanh.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qtanh.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [557/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qsoftmax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qsoftmax.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qsoftmax.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [558/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qthreshold.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qthreshold.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [559/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/library.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/library.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [560/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/qconv_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qconv_unpack.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qconv_unpack.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [561/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/LinearAlgebra.cpp.o [562/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/SparseBlasImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SparseBlasImpl.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [563/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/SparseCsrLinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SparseCsrLinearAlgebra.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SparseCsrLinearAlgebra.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [564/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/qlinear_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qlinear_unpack.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qlinear_unpack.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [565/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/BinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/BinaryOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [566/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Conv.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Conv.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [567/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/ConvPrepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Common.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/ConvPrepack.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [568/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Copy.cpp.o [569/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/SpectralOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SpectralOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [570/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/IDeepRegistration.cpp.o [571/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Gelu.cpp.o [572/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Linear.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [573/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/MKLDNNCommon.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/MKLDNNCommon.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [574/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Matmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Matmul.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Matmul.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [575/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/MKLDNNConversions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/MKLDNNConversions.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/MKLDNNConversions.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [576/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/MkldnnTensorMath.cpp.o [577/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Normalization.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [578/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Pooling.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [579/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/OpContext.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/OpContext.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/OpContext.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [580/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Prelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Prelu.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [581/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/RegisterMkldnnOpContextClass.cpp.o [582/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Relu.cpp.o [583/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/SoftMax.cpp.o [584/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/TensorFactories.cpp.o [585/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/RNN.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/RNN.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/RNN.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [586/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/TensorShape.cpp.o [587/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/UnaryOps.cpp.o [588/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [589/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/transformers/attention.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/attention.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [590/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/transformers/sdp_utils_cpp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [591/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/utils/Factory.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/utils/Factory.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [592/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/transformers/transformer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/transformer.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [593/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Activation.cpp.o [594/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/AveragePooling.cpp.o [595/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/ChannelShuffle.cpp.o [596/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Init.cpp.o [597/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Convolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Convolution.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Convolution.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [598/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Linear.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Linear.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [599/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/MaxPooling.cpp.o [600/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/OpContext.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Convolution.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [601/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Shim.cpp.o [602/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/RegisterOpContextClass.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/RegisterOpContextClass.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [603/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/CompositeViewCopyKernels.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/CompositeViewCopyKernels.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [604/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [605/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_0.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [606/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_1.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [607/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_3.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_3.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [608/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_2.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [609/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterBackendSelect.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterBackendSelect.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [610/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_4.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_4.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [611/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [612/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeExplicitAutograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutograd.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutograd.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [613/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeExplicitAutogradNonFunctional.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutogradNonFunctional.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutogradNonFunctional.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [614/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeImplicitAutogradNestedTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutogradNestedTensor.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutogradNestedTensor.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [615/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeImplicitAutograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutograd.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutograd.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [616/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_0.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [617/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_1.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [618/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_2.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [619/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_3.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_3.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [620/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterMkldnnCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMkldnnCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMkldnnCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [621/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterNestedTensorCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [622/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterNestedTensorMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [623/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterQuantizedCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [624/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [625/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterQuantizedMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [626/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSchema.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSchema.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [627/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [628/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseCsrCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [629/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseCsrMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [630/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [631/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterZeroTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterZeroTensor.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterZeroTensor.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [632/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/UfuncCPU_add.cpp.o [633/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/ATenOpList.cpp.o [634/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/quantized/QTensorImpl.cpp.o [635/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/TensorMethods.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorMethods.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [636/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_bind.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [637/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/quantized/Quantizer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/quantized/Quantizer.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [638/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_model_loader.cpp.o [639/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_wrapper.cpp.o [640/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/UfuncCPUKernel_add.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/AccumulateType.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Math.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec_base.h:27, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional_base.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ufunc/add.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/UfuncCPUKernel_add.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/UfuncCPUKernel_add.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [641/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_register.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_register.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [642/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/spherical_bessel_j0.cpp.DEFAULT.cpp.o [643/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k1.cpp.DEFAULT.cpp.o [644/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k0.cpp.DEFAULT.cpp.o [645/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [646/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int8mm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int8mm_kernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int8mm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [647/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [648/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int4mm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int4mm_kernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int4mm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [649/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/batch_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [650/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/airy_ai.cpp.DEFAULT.cpp.o [651/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/group_norm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/group_norm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [652/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/WeightNormKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/WeightNormKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/WeightNormKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [653/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [654/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [655/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [656/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Unfold2d.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/Unfold2d.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/Unfold2d.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [657/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [658/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SumKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/AccumulateType.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SumKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SumKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [659/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/StackKernel.cpp.DEFAULT.cpp.o [660/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [661/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Generator.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnaryOps.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [662/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SparseFactories.cpp.DEFAULT.cpp.o [663/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SoftMaxKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [664/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SortingKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SortingKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SortingKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [665/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NonEmptyUtils.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [666/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RenormKernel.cpp.DEFAULT.cpp.o [667/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [668/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOpsUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [669/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [670/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [671/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PowKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PowKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PowKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [672/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PixelShuffleKernel.cpp.DEFAULT.cpp.o [673/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [674/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [675/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MultinomialKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MultinomialKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [676/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxUnpoolKernel.cpp.DEFAULT.cpp.o [677/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPooling.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MaxPooling.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MaxPooling.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [678/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PaddingKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PaddingKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PaddingKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [679/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptivePooling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MaxPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [680/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [681/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LerpKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/OpMathType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Lerp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/LerpKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/LerpKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [682/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/HistogramKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/aminmax.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/HistogramKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Histogram.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [683/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/IndexKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/IndexKernel.cpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/IndexKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [684/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedSGDKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [685/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedAdamKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [686/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [687/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [688/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/GridSamplerKernel.cpp.DEFAULT.cpp.o [689/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FillKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch_v2.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FillKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FillKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [690/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistributionKernels.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DistributionKernels.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [691/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistanceOpsKernel.cpp.DEFAULT.cpp.o [692/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [693/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CrossKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CrossKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CrossKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [694/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [695/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ComplexKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ComplexKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ComplexKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [696/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ChannelShuffleKernel.cpp.DEFAULT.cpp.o [697/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CatKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CatKernel.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CatKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CatKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [698/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/BlasKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/BlasKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/BlasKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [699/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CopyKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CopyKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CopyKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [700/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AvgPoolKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AvgPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AvgPoolKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [701/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [702/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [703/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BinaryOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [704/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [705/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/vulkan/Context.cpp.o [706/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/metal/Context.cpp.o [707/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/core/common.cc.o [708/1912] Building C object caffe2/CMakeFiles/torch_cpu.dir/__/third_party/miniz-2.1.0/miniz.c.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/miniz-2.1.0/miniz.c:3157:9: note: ‘#pragma message: Using fopen, ftello, fseeko, stat() etc. path for file I/O - this path may not support large files.’ 3157 | #pragma message("Using fopen, ftello, fseeko, stat() etc. path for file I/O - this path may not support large files.") | ^~~~~~~ [709/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/serialize/inline_container.cc.o [710/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/serialize/istream_adapter.cc.o [711/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/serialize/file_adapter.cc.o [712/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/serialize/crc.cc.o [713/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/serialize/read_adapter_interface.cc.o [714/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/utils/string_utils.cc.o [715/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/utils/threadpool/ThreadPool.cc.o [716/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/utils/threadpool/pthreadpool-cpp.cc.o [717/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/utils/threadpool/thread_pool_guard.cpp.o [718/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/utils/proto_wrap.cc.o [719/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/adagrad.cc.o [720/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/batch_box_cox.cc.o [721/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/embedding_lookup.cc.o [722/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/embedding_lookup_idx.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/caffe2/perfkernels/embedding_lookup_idx.cc:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [723/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/fused_8bit_rowwise_embedding_lookup.cc.o [724/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/fused_8bit_rowwise_embedding_lookup_idx.cc.o [725/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/fused_nbit_rowwise_conversion.cc.o [726/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/lstm_unit_cpu_common.cc.o [727/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/math_cpu_base.cc.o [728/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/perfkernels/typed_axpy.cc.o [729/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Activation.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/Activation.cpp:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/Activation.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [730/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/ViewFuncs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [731/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/Functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/Functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [732/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/VariableType_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_0.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [733/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/VariableType_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_1.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [734/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/VariableType_2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [735/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/VariableType_3.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_3.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [736/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/VariableType_4.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_4.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [737/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/TraceType_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_0.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_0.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [738/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/TraceType_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_1.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_1.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [739/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/TraceType_2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_2.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_2.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [740/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/TraceType_3.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_3.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_3.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [741/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/TraceType_4.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_4.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_4.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [742/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/ADInplaceOrViewType_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ADInplaceOrViewType_0.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [743/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [744/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/generated/ADInplaceOrViewType_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ADInplaceOrViewType_1.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [745/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/generated/RegisterAutogradLazy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterAutogradLazy.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterAutogradLazy.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [746/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/generated/LazyNativeFunctions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNativeFunctions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h: In function ‘std::optional torch::lazy::GetBackendDevice(const T&, const Args& ...) [with T = std::optional; Args = {}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h:89:30: warning: infinite recursion detected [-Winfinite-recursion] 89 | std::optional GetBackendDevice( | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h:92:8: note: recursive call 92 | auto optional_device = GetBackendDevice(tensor); | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h: In function ‘std::optional torch::lazy::GetBackendDevice(const T&, const Args& ...) [with T = std::optional; Args = {}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h:89:30: warning: infinite recursion detected [-Winfinite-recursion] 89 | std::optional GetBackendDevice( | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h:92:8: note: recursive call 92 | auto optional_device = GetBackendDevice(tensor); | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h: In function ‘std::optional torch::lazy::GetBackendDevice(const T&, const Args& ...) [with T = c10::SymInt; Args = {}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h:89:30: warning: infinite recursion detected [-Winfinite-recursion] 89 | std::optional GetBackendDevice( | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.h:92:8: note: recursive call 92 | auto optional_device = GetBackendDevice(tensor); | ^~~~~~~~~~~~~~~ [747/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/anomaly_mode.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/anomaly_mode.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [748/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/autograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [749/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/generated/RegisterLazy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterLazy.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterLazy.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [750/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/autograd_meta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_meta.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [751/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/autograd_not_implemented_fallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [752/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/cpp_hook.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/cpp_hook.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [753/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/engine.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [754/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/forward_grad.cpp.o [755/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/custom_function.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [756/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/function.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [757/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/functions/accumulate_grad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CachedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [758/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/functions/basic_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/basic_ops.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/basic_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [759/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/functions/tensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/tensor.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/tensor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [760/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/functions/utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/utils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [761/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/input_metadata.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_metadata.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_metadata.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [762/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/jit_decomp_interface.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/jit_decomp_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/jit_decomp_interface.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [763/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/input_buffer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_buffer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/input_buffer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [764/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/profiler_kineto.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [765/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/profiler_legacy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_legacy.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [766/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/saved_variable.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/saved_variable.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [767/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/utils/warnings.cpp.o [768/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/record_function_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [769/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/variable_info.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable_info.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [770/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/aoti_runner/model_container_runner.cpp.o [771/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/aoti_runner/model_container_runner_cpu.cpp.o [772/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/aoti_torch/mkldnn_tensor.cpp.o [773/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/aoti_torch/shim_common.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/proxy_executor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/shim_common.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [774/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/variable.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [775/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/aoti_torch/tensor_converter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/tensor_converter.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [776/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/inductor/inductor_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/mm.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/inductor_ops.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/inductor_ops.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [777/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/api/function_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [778/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/api/object.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [779/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/api/module.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [780/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/backends/backend_debug_handler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [781/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/backends/backend_debug_info.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_info.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [782/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/backends/backend_detail.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [783/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/backends/backend_interface.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [784/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/backends/backend_resolver.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_resolver.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_resolver.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [785/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/codegen.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/arg_spec.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/codegen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [786/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/compiler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [787/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/executor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_spec.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/executor.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [788/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/fallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_spec.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/fallback.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [789/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/interface.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [790/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/kernel_cache.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_spec.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_cache.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/kernel_spec.h:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [791/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/builtin_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/builtin_functions.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/builtin_functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [792/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/canonicalize_modified_loop.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/canonicalize_modified_loop.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [793/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/edit_distance.cpp.o [794/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/convert_to_ssa.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/convert_to_ssa.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [795/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/exit_transforms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/exit_transforms.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/exit_transforms.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [796/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/inline_loop_condition.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/inline_loop_condition.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/inline_loop_condition.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [797/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/name_mangler.cpp.o [798/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [799/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/schema_matching.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_matching.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_matching.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [800/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/ir_emitter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp: In member function ‘std::vector torch::jit::to_ir::emitFormalArguments(const torch::jit::Def&, const torch::jit::Self*, const c10::FunctionSchema&, torch::jit::Block*)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:853:19: warning: possibly dangling reference to a temporary [-Wdangling-reference] 853 | const auto& name = (*it).ident().name(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:853:44: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Param::ident() const().torch::jit::Ident::name()’ 853 | const auto& name = (*it).ident().name(); | ~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:871:13: warning: possibly dangling reference to a temporary [-Wdangling-reference] 871 | auto& name = (*it).ident().name(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:871:38: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Param::ident() const().torch::jit::Ident::name()’ 871 | auto& name = (*it).ident().name(); | ~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp: In member function ‘void torch::jit::to_ir::emitDelete(const torch::jit::Delete&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:1013:28: warning: possibly dangling reference to a temporary [-Wdangling-reference] 1013 | const SourceRange& val_range = subscript.value().range(); | ^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:1013:63: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Subscript::value() const().torch::jit::Expr::torch::jit::TreeView.torch::jit::TreeView::range()’ 1013 | const SourceRange& val_range = subscript.value().range(); | ~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp: In member function ‘torch::jit::RefinementSet torch::jit::to_ir::findIsNoneRefinements(const torch::jit::Expr&, torch::jit::Value*, const torch::jit::Expr&, torch::jit::Value*, int)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:1181:24: warning: possibly dangling reference to a temporary [-Wdangling-reference] 1181 | const std::string& name = Var(lhs).name().name(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:1181:51: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Var::name() const().torch::jit::Ident::name()’ 1181 | const std::string& name = Var(lhs).name().name(); | ~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp: In member function ‘torch::jit::CondValue torch::jit::to_ir::emitHasAttr(const torch::jit::Expr&, const torch::jit::Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:2146:24: warning: possibly dangling reference to a temporary [-Wdangling-reference] 2146 | const std::string& name = StringLiteral(attrExpr).text(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:2146:59: note: the temporary was destroyed at the end of the full expression ‘torch::jit::StringLiteral(torch::jit::TreeView::operator torch::jit::TreeRef() const()).torch::jit::StringLiteral::text()’ 2146 | const std::string& name = StringLiteral(attrExpr).text(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp: In member function ‘std::shared_ptr torch::jit::to_ir::emitApplySpecialForm(c10::Symbol, torch::jit::Apply&, std::shared_ptr, const c10::TypePtr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:3457:28: warning: possibly dangling reference to a temporary [-Wdangling-reference] 3457 | const std::string& name = StringLiteral(selector).text(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:3457:63: note: the temporary was destroyed at the end of the full expression ‘torch::jit::StringLiteral(torch::jit::TreeView::operator torch::jit::TreeRef() const()).torch::jit::StringLiteral::text()’ 3457 | const std::string& name = StringLiteral(selector).text(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp: In member function ‘std::shared_ptr torch::jit::to_ir::emitSubscript(const torch::jit::Subscript&, c10::TypePtr)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:5213:24: warning: possibly dangling reference to a temporary [-Wdangling-reference] 5213 | const SourceRange& val_range = subscript.value().range(); | ^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.cpp:5213:59: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Subscript::value() const().torch::jit::Expr::torch::jit::TreeView.torch::jit::TreeView::range()’ 5213 | const SourceRange& val_range = subscript.value().range(); | ~~~~~~~~~~~~~~~~~~~~~~~^~ [801/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/script_type_parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp: In member function ‘std::optional, int> > torch::jit::ScriptTypeParser::parseBroadcastList(const torch::jit::Expr&) const’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:126:11: warning: possibly dangling reference to a temporary [-Wdangling-reference] 126 | auto& name = var.name().name(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:126:33: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Var::name() const().torch::jit::Ident::name()’ 126 | auto& name = var.name().name(); | ~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp: In member function ‘std::optional > torch::jit::ScriptTypeParser::parseBaseTypeName(const torch::jit::Expr&) const’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:208:26: warning: possibly dangling reference to a temporary [-Wdangling-reference] 208 | const std::string& name = select.selector().name(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:208:55: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Select::selector() const().torch::jit::Ident::name()’ 208 | const std::string& name = select.selector().name(); | ~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp: In member function ‘c10::TypePtr torch::jit::ScriptTypeParser::parseTypeFromExprImpl(const torch::jit::Expr&) const’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:264:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 264 | const auto& type_name = StringLiteral(expr).text(); | ^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.cpp:264:53: note: the temporary was destroyed at the end of the full expression ‘torch::jit::StringLiteral(torch::jit::TreeView::operator torch::jit::TreeRef() const()).torch::jit::StringLiteral::text()’ 264 | const auto& type_name = StringLiteral(expr).text(); | ~~~~~~~~~~~~~~~~~~~~~~~~^~ [802/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/sugared_value.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [803/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/tracer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [804/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/versioned_symbols.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/versioned_symbols.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/versioned_symbols.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [805/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/attributes.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/attributes.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [806/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/alias_analysis.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [807/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/graph_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [808/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/constants.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [809/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/ir.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [810/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/irparser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/irparser.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [811/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/node_hashing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/node_hashing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [812/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/scope.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/scope.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [813/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/type_hashing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [814/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/ir/subgraph_matcher.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/subgraph_matcher.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/subgraph_matcher.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [815/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/jit_log.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_log.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_log.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [816/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/jit_opt_limit.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_opt_limit.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/jit_opt_limit.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [817/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/compatibility/runtime_compatibility.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/runtime_compatibility.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [818/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/compatibility/model_compatibility.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/model_compatibility.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [819/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/function.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/function.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [820/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/flatbuffer_loader.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.cpp:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/flatbuffer_loader.cpp:36: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [821/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/import.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.cpp:24: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [822/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/interpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/code.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/interpreter.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/interpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [823/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/module.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [824/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/nnc/backend.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/backend.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [825/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/nnc/aot_compiler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/aot_compiler.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/aot_compiler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [826/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/nnc/registry.cpp.o [827/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/observer.cpp.o [828/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/nnc/context.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp: In constructor ‘torch::jit::mobile::nnc::Function::Function(const c10::IValue&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:137:53: warning: possibly dangling reference to a temporary [-Wdangling-reference] 137 | dict.at("input_specs").toTupleRef().elements()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:137:52: note: the temporary was destroyed at the end of the full expression ‘(& c10::Dict::at(const Key&) const [with Key = c10::IValue; Value = c10::IValue](c10::IValue(((const char*)"input_specs"))).c10::IValue::toTupleRef())->c10::ivalue::Tuple::elements()’ 137 | dict.at("input_specs").toTupleRef().elements()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:143:54: warning: possibly dangling reference to a temporary [-Wdangling-reference] 143 | dict.at("output_specs").toTupleRef().elements()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:143:53: note: the temporary was destroyed at the end of the full expression ‘(& c10::Dict::at(const Key&) const [with Key = c10::IValue; Value = c10::IValue](c10::IValue(((const char*)"output_specs"))).c10::IValue::toTupleRef())->c10::ivalue::Tuple::elements()’ 143 | dict.at("output_specs").toTupleRef().elements()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:152:55: warning: possibly dangling reference to a temporary [-Wdangling-reference] 152 | dict.at("sym_shape_pos").toTupleRef().elements()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/nnc/context.cpp:152:54: note: the temporary was destroyed at the end of the full expression ‘(& c10::Dict::at(const Key&) const [with Key = c10::IValue; Value = c10::IValue](c10::IValue(((const char*)"sym_shape_pos"))).c10::IValue::toTupleRef())->c10::ivalue::Tuple::elements()’ 152 | dict.at("sym_shape_pos").toTupleRef().elements()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ [829/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/parse_bytecode.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_bytecode.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [830/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/parse_operators.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/parse_operators.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [831/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/prim_ops_registery.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/prim_ops_registery.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/prim_ops_registery.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [832/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/promoted_prim_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/promoted_prim_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [833/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/quantization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.cpp: In member function ‘void torch::jit::mobile::quantization::PTQQuanizationHelper::quantize_dynamic(torch::jit::mobile::Module&, const std::string&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.cpp:55:19: warning: possibly dangling reference to a temporary [-Wdangling-reference] 55 | const Function& to_be_copied = | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/quantization.cpp:56:60: note: the temporary was destroyed at the end of the full expression ‘(&(& m)->torch::jit::mobile::Module::find_method(quantized_method_name).std::optional::value())->torch::jit::mobile::Method::function()’ 56 | m.find_method(quantized_method_name).value().function(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ [834/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/register_ops_common_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/register_ops_common_utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/register_ops_common_utils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [835/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/type_parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/type_parser.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [836/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/upgrader_mobile.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/upgrader_mobile.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/upgrader_mobile.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [837/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/operator_upgraders/upgraders.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [838/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/operator_upgraders/utils.cpp.o [839/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/operator_upgraders/version_map.cpp.o [840/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/operator_upgraders/upgraders_entry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/operator_upgraders/upgraders_entry.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [841/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/add_if_then_else.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/add_if_then_else.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/add_if_then_else.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [842/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/annotate_warns.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/annotate_warns.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/annotate_warns.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [843/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/bailout_graph.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/bailout_graph.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [844/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/canonicalize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [845/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/batch_mm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/batch_mm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/batch_mm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [846/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [847/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/check_strict_fusion.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/check_strict_fusion.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/check_strict_fusion.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [848/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/clear_profiling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_profiling.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [849/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/clear_undefinedness.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/clear_undefinedness.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [850/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/common_subexpression_elimination.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/common_subexpression_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/common_subexpression_elimination.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [851/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/concat_opt.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/concat_opt.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/concat_opt.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [852/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/constant_pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [853/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/constant_propagation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_propagation.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_propagation.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [854/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/create_autodiff_subgraphs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_autodiff_subgraphs.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_autodiff_subgraphs.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [855/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/create_functional_graphs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_functional_graphs.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/create_functional_graphs.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [856/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dbr_quantization/remove_redundant_aliases.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [857/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/dead_code_elimination.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [858/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/decompose_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/decompose_ops.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [859/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/device_type_analysis.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/device_type_analysis.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [860/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/dtype_analysis.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dtype_analysis.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dtype_analysis.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [861/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/eliminate_no_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/eliminate_no_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/eliminate_no_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [862/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/erase_number_types.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/erase_number_types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/erase_number_types.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [863/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/fixup_trace_scope_blocks.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fixup_trace_scope_blocks.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fixup_trace_scope_blocks.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [864/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/fold_conv_bn.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_conv_bn.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_conv_bn.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [865/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/fold_linear_bn.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_linear_bn.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fold_linear_bn.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [866/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_concat_linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_concat_linear.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [867/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [868/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_conv_folding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_folding.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_folding.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [869/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/freeze_module.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/freeze_module.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [870/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_graph_optimizations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_graph_optimizations.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [871/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_linear_folding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_folding.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_folding.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [872/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_linear_transpose.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_linear_transpose.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [873/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/fuse_linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_linear.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_linear.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [874/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/frozen_ops_to_mkldnn.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_ops_to_mkldnn.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_ops_to_mkldnn.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [875/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/fuse_relu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_relu.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/fuse_relu.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [876/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/graph_rewrite_helper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_rewrite_helper.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_rewrite_helper.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [877/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/graph_fuser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/graph_fuser.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [878/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/hoist_conv_packed_params.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/hoist_conv_packed_params.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [879/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/guard_elimination.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/guard_elimination.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [880/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/inline_autodiff_subgraphs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_autodiff_subgraphs.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_autodiff_subgraphs.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [881/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/inline_fork_wait.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_fork_wait.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_fork_wait.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [882/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/inliner.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [883/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/inline_forked_closures.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inline_forked_closures.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [884/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/inplace_check.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inplace_check.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inplace_check.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [885/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/insert_guards.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/insert_guards.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [886/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/integer_value_refinement.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/integer_value_refinement.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [887/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/lift_closures.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lift_closures.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [888/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/liveness.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/liveness.h:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [889/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/loop_unrolling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/loop_unrolling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/loop_unrolling.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [890/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/lower_grad_of.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_grad_of.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_grad_of.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [891/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/lower_tuples.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_tuples.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_tuples.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [892/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/metal_rewrite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/metal_rewrite.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/metal_rewrite.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [893/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/mkldnn_rewrite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/mkldnn_rewrite.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [894/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/normalize_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/normalize_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/normalize_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [895/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/pass_manager.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/pass_manager.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/pass_manager.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [896/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/peephole_alias_sensitive.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_alias_sensitive.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [897/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/peephole.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [898/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/peephole_dict_idioms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_dict_idioms.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [899/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/peephole_list_idioms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_list_idioms.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [900/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/peephole_non_tensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/peephole_non_tensor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [901/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/prepack_folding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/prepack_folding.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [902/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/dedup_module_uses.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/dedup_module_uses.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/dedup_module_uses.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [903/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/fusion_passes.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/fusion_passes.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/fusion_passes.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [904/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/finalize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/finalize.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/finalize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [905/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/helper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.cpp: In function ‘void torch::jit::cloneMethod(Module&, const std::string&, const std::string&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.cpp:335:19: warning: possibly dangling reference to a temporary [-Wdangling-reference] 335 | const Function& method = module.get_method(orig_method_name).function(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/helper.cpp:335:72: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Object::get_method(const std::string&) const((* & orig_method_name)).torch::jit::Method::function()’ 335 | const Function& method = module.get_method(orig_method_name).function(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ [906/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/insert_observers.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_observers.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_observers.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [907/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/quantization_type.cpp.o [908/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/insert_quant_dequant.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_quant_dequant.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/insert_quant_dequant.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [909/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/refine_tuple_types.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/refine_tuple_types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/refine_tuple_types.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [910/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/quantization/register_packed_params.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/register_packed_params.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/quantization/register_packed_params.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [911/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/remove_dropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_dropout.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_dropout.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [912/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/remove_exceptions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/constant_pooling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_exceptions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [913/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/remove_expands.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_expands.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_expands.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [914/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/remove_mutation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [915/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/remove_redundant_profiles.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_redundant_profiles.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [916/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/replacement_of_old_operators.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/replacement_of_old_operators.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/replacement_of_old_operators.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [917/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/requires_grad_analysis.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/requires_grad_analysis.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/requires_grad_analysis.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [918/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/restore_mutation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_mutation.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/restore_mutation.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [919/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/shape_analysis.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/shape_analysis.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [920/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/subgraph_rewrite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/subgraph_rewrite.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/subgraph_rewrite.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [921/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/specialize_autogradzero.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/specialize_autogradzero.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/specialize_autogradzero.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [922/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/symbolic_shape_cache.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_cache.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [923/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/symbolic_shape_analysis.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [924/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [925/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/tensorexpr_fuser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/tensorexpr_fuser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/tensorexpr_fuser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [926/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [927/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/utils/memory_dag.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/memory_dag.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/memory_dag.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [928/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/utils/op_registry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/op_registry.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/op_registry.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [929/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/utils/optimization_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/optimization_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/optimization_utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [930/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/utils/subgraph_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/subgraph_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/subgraph_utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [931/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/value_refinement_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/alias_analysis.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/value_refinement_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/value_refinement_utils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [932/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/variadic_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/variadic_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/variadic_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [933/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/vulkan_rewrite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/vulkan_rewrite.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/vulkan_rewrite.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [934/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/python/update_graph_executor_opt.cpp.o [935/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/python/utf8_decoding_ignore.cpp.o [936/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/xnnpack_rewrite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/xnnpack_rewrite.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [937/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/argument_spec.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/argument_spec.h:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [938/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/autodiff.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/autodiff.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/autodiff.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [939/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/decomposition_registry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [940/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/decomposition_registry_util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/decomposition_registry_util.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [941/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/instruction.cpp.o [942/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/graph_executor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp:37: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp: In member function ‘void torch::jit::{anonymous}::DifferentiableGraphBackward::addOutputForIValue(const c10::IValue&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp:333:30: warning: possibly dangling reference to a temporary [-Wdangling-reference] 333 | for (const at::Tensor& tensor : value.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp:333:58: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 333 | for (const at::Tensor& tensor : value.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp: In member function ‘void torch::jit::{anonymous}::DifferentiableGraphBackward::addInputIValue(const c10::IValue&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp:364:30: warning: possibly dangling reference to a temporary [-Wdangling-reference] 364 | for (const at::Tensor& tensor : tensors) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor.cpp:364:39: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 364 | for (const at::Tensor& tensor : tensors) { | ^~~~~~~ [943/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/interpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [944/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/interpreter/preprocess_graph.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/preprocess_graph.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/preprocess_graph.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [945/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/jit_exception.cpp.o [946/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/interpreter/frame.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/code_impl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/frame.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter/frame.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [947/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/logging.cpp.o [948/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/jit_trace.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_trace.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir_views.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/jit_trace.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [949/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/print_handler.cpp.o [950/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/operator.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [951/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/profiling_record.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_record.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [952/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/profiling_graph_executor_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [953/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/script_profile.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/script_profile.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [954/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/register_ops_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [955/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/serialized_shape_function_registry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/serialized_shape_function_registry.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [956/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/slice_indices_adjust.cpp.o [957/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/fusion.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/fusion.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/fusion.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [958/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/simple_graph_executor_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/simple_graph_executor_impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/graph_executor_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/profiling_graph_executor_impl.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [959/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [960/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/memory_planner.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/memory_planner.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/memory_planner.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/impl.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [961/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/generated_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/generated_ops.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [962/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/native_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/inliner.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp:1178:15: warning: possibly dangling reference to a temporary [-Wdangling-reference] 1178 | auto& elems = future->value().toTupleRef().elements(); | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h:46:28: note: in definition of macro ‘REGISTER_NATIVE_OPERATOR_FUNCTOR’ 46 | const SROpFunctor fn = __VA_ARGS__; \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/native_ops.cpp:1178:60: note: the temporary was destroyed at the end of the full expression ‘(& c10::ivalue::Future::value()().c10::IValue::toTupleRef())->c10::ivalue::Tuple::elements()’ 1178 | auto& elems = future->value().toTupleRef().elements(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h:46:28: note: in definition of macro ‘REGISTER_NATIVE_OPERATOR_FUNCTOR’ 46 | const SROpFunctor fn = __VA_ARGS__; \ | ^~~~~~~~~~~ [963/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/passes.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/passes.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/passes.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [964/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/EmptyTensor.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/ops.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [965/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/symbolic_script.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_script.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [966/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/static/te_wrapper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/te_wrapper.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [967/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/symbolic_shape_registry_util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry_util.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [968/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/symbolic_shape_registry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/symbolic_shape_registry.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [969/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/vararg_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/vararg_functions.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/vararg_functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [970/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/callstack_debug_info_serialization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/callstack_debug_info_serialization.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/callstack_debug_info_serialization.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [971/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/import_export_helpers.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_export_helpers.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [972/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/import_read.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_read.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [973/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/import.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [974/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/pickle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [975/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/import_source.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.cpp: In member function ‘void torch::jit::SourceImporterImpl::importNamedType(const std::string&, const torch::jit::ClassDef&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.cpp:286:15: warning: possibly dangling reference to a temporary [-Wdangling-reference] 286 | const auto& superclass_name = Var(class_def.superclass().get()).name().name(); | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.cpp:286:78: note: the temporary was destroyed at the end of the full expression ‘torch::jit::Var::name() const().torch::jit::Ident::name()’ 286 | const auto& superclass_name = Var(class_def.superclass().get()).name().name(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ [976/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/python_print.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/python_print.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/python_print.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [977/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/pickler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickler.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [978/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/source_range_serialization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [979/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/type_name_uniquer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/type_hashing.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/type_name_uniquer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/type_name_uniquer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [980/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/block_codegen.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/block_codegen.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/block_codegen.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [981/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/unpickler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp: In function ‘void torch::jit::restoreAccurateTypeTags(const c10::IValue&, const c10::TypePtr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp:145:28: warning: possibly dangling reference to a temporary [-Wdangling-reference] 145 | for (const IValue& item : lst) { | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.cpp:145:35: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 145 | for (const IValue& item : lst) { | ^~~ [982/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/bounds_inference.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_inference.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_inference.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [983/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/bounds_overlap.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [984/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/codegen.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/analysis.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [985/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/cpp_codegen.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_codegen.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cpp_codegen.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [986/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/expr.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [987/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/eval.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [988/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/external_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/import_source.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [989/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/external_functions_codegen.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_codegen.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [990/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/external_functions_registry.cpp.o [991/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/external_functions_core.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_core.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/external_functions_core.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [992/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/hash_provider.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [993/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/intrinsic_symbols.cpp.o [994/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/graph_opt.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/graph_opt.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/graph_opt.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [995/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/exceptions.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [996/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir_cloner.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_cloner.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [997/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir_mutator.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/codegen.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/eval.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_mutator.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [998/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir_printer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_printer.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_printer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_printer.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [999/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir_verifier.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_verifier.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1000/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir_simplifier.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1001/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/ir_visitor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_visitor.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1002/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/llvm_codegen.cpp.o [1003/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/llvm_jit.cpp.o [1004/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/kernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp: In member function ‘torch::jit::tensorexpr::ExprHandle torch::jit::tensorexpr::TensorExprKernel::getVarForShape(const c10::ShapeSymbol&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp:485:21: warning: redundant move in return statement [-Wredundant-move] 485 | return std::move(var); | ~~~~~~~~~^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp:485:21: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp: In member function ‘torch::jit::tensorexpr::ExprHandle torch::jit::tensorexpr::TensorExprKernel::getStrideArg(size_t, size_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp:1027:21: warning: redundant move in return statement [-Wredundant-move] 1027 | return std::move(var); | ~~~~~~~~~^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.cpp:1027:21: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/attributes.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1005/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/loopnest.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/analysis.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1006/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/loopnest_randomization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest_randomization.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1007/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/lowerings.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/jit_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1008/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/mem_dependency_checker.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/mem_dependency_checker.h:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1009/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/conv2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/loopnest.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/conv2d.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1010/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/matmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/matmul.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1011/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/misc.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/fwd_decls.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/expr.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1012/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/norm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/norm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1013/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/quantization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/bounds_overlap.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir_simplifier.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/quantization.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/quantization.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1014/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/pointwise.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/lowerings.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/misc.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/pointwise.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1015/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/softmax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/softmax.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/softmax.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1016/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/operators/reduction.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/reduction.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/operators/reduction.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1017/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/reduction.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1018/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/tensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/reduction.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensor.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1019/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/types.cpp.o [1020/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/registerizer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/hash_provider.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/registerizer.h:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1021/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/tensorexpr/unique_name_manager.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/ir.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/unique_name_manager.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1022/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/testing/hooks_for_testing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/hooks_for_testing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/hooks_for_testing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1023/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/testing/file_check.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/testing/file_check.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1024/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/backend/backend_device.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_device.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1025/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/backend/backend_interface.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1026/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/config.cpp.o [1027/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/backend/lowering_context.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1028/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/hash.cpp.o [1029/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/debug_util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1030/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/helpers.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/helpers.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/helpers.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1031/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/ir.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1032/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/ir_metadata.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_metadata.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1033/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/ir_dump_util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_dump_util.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_dump_util.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1034/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/ir_util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_util.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir_util.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1035/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/metrics.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/metrics.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1036/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/multi_wait.cpp.o [1037/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/lazy_graph_executor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/lowering_context.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/lazy_graph_executor.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/lazy_graph_executor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1038/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/ops/arithmetic_ir_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/arithmetic_ir_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/arithmetic_ir_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1039/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/ops/utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ops/utils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1040/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/permutation_util.cpp.o [1041/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/shape.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1042/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/tensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1043/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/shape_inference.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape_inference.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape_inference.cpp:50: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1044/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/tensor_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_impl.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1045/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/thread_pool.cpp.o [1046/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/tensor_util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor_util.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1047/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/monitor/counters.cpp.o [1048/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/monitor/events.cpp.o [1049/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/core/trie.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/trie.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/trie.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1050/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/combined_traceback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1051/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/collection.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1052/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/kineto_client_interface.cpp.o [1053/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/data_flow.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/data_flow.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1054/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/kineto_shim.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1055/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/orchestration/observer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1056/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/orchestration/python_tracer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/python_tracer.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/python_tracer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1057/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/perf.cpp.o [1058/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/orchestration/vulkan.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/vulkan.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/vulkan.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1059/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/standalone/itt_observer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/itt_observer.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/itt_observer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1060/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/standalone/execution_trace_observer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/execution_trace_observer.cpp:25: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1061/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/standalone/nvtx_observer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.cpp: In function ‘std::__cxx11::list > torch::profiler::impl::flattenOpIdList(const c10::List&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.cpp:73:27: warning: possibly dangling reference to a temporary [-Wdangling-reference] 73 | for (const c10::IValue& input : list) { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/nvtx_observer.cpp:73:35: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 73 | for (const c10::IValue& input : list) { | ^~~~ [1062/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/stubs/base.cpp.o [1063/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/standalone/privateuse1_observer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/privateuse1_observer.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/standalone/privateuse1_observer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1064/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/unwind/unwind_fb.cpp.o [1065/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/unwind/unwind.cpp.o [1066/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/utils/cpp_stacktraces.cpp.o [1067/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/profiler/util.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.cpp: In function ‘std::vector > torch::profiler::impl::flattenList(const c10::List&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.cpp:166:27: warning: possibly dangling reference to a temporary [-Wdangling-reference] 166 | for (const c10::IValue& input : list) { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.cpp:166:35: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 166 | for (const c10::IValue& input : list) { | ^~~~ [1068/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/utils/schema_info.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/schema_info.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1069/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/utils/variadic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1070/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/utils/tensor_flatten.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_flatten.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_flatten.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1071/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/cuda/interface.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/cuda/interface.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1072/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/autocast.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/autocast.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/autocast.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1073/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/lower_graph.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_graph.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/lower_graph.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1074/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/remove_inplace_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_inplace_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/remove_inplace_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1075/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/passes/utils/check_alias_annotation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.cpp: In function ‘c10::IValue torch::jit::{anonymous}::deepCopy(const c10::IValue&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.cpp:31:24: warning: possibly dangling reference to a temporary [-Wdangling-reference] 31 | for (const IValue& value : source) { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/utils/check_alias_annotation.cpp:31:32: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 31 | for (const IValue& value : source) { | ^~~~~~ [1076/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/register_c10_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_c10_ops.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1077/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/register_prim_ops_fulljit.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops_fulljit.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops_fulljit.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1078/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/register_prim_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h: In instantiation of ‘void torch::jit::listContains(Stack&) [with T = std::__cxx11::basic_string; Stack = std::vector]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_prim_ops.cpp:1100:34: required from here 1100 | aliasAnalysisFromSchema()), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:333:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 333 | for (const T& item : list) { | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_ops_utils.h:333:3: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::operator std::conditional_t&, std::__cxx11::basic_string >()’ 333 | for (const T& item : list) { | ^~~ [1079/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/debug_info.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1080/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/runtime/register_special_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_special_ops.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1081/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/config.cpp.o [1082/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/dynamic_ir.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/ir.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/dynamic_ir.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/dynamic_ir.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/dynamic_ir.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1083/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ops/device_data.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/device_data.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/device_data.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1084/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ops/generic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/generic.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ops/generic.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1085/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/tensor_aten_ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/tensor_aten_ops.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/tensor_aten_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1086/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_autograd_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_autograd_functions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1087/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_backend_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1088/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_eager_fallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_eager_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_eager_fallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1089/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_lowering_context.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_interface.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_backend_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_lowering_context.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1090/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_node.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1091/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_native_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_native_functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1092/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/lazy/ts_backend/ts_node_lowering.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/jit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node_lowering.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/ts_backend/ts_node_lowering.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1093/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/train/export_data.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/export_data.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/export_data.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1094/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/import_data.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import_data.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1095/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/train/random.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/random.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/random.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1096/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/train/optim/sgd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/optim/sgd.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/optim/sgd.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1097/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/train/sequential.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/sequential.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/train/sequential.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1098/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/flatbuffer_serializer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/flatbuffer_serializer.h:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1099/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/utils/out_types.cpp.o [1100/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/TraceTypeManual.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/op_registration.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/TraceTypeManual.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/TraceTypeManual.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1101/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/FunctionsManual.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/FunctionsManual.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1102/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/jit.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/jit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/jit.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/jit.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1103/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/compatibility/backport.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1104/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/autograd/VariableTypeManual.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RedispatchFunctions.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeManual.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/VariableTypeManual.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1105/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/onnx.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/onnx.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/onnx.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1106/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/mobile/compatibility/backport_manager.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport_manager.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/compilation_unit.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/debug_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/import.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/mobile/compatibility/backport_manager.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1107/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/export_bytecode.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_detail.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_debug_handler.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_bytecode.cpp:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1108/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/export.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1109/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/serialization/export_module.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/export_module.cpp:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1110/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cpu/fused_kernel.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1111/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/utils/byte_order.cpp.o [1112/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/api/module_save.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module_save.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1113/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/control_collectives/StoreCollectives.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/control_collectives/StoreCollectives.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1114/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/Backend.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1115/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/FileStore.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.hpp:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1116/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/GlooDeviceFactory.cpp.o [1117/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/GroupRegistry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1118/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/Functional.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Functional.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/GroupRegistry.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Functional.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1119/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/ParamCommsUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ParamCommsUtils.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ParamCommsUtils.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1120/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/PrefixStore.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1121/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/Ops.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1122/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/ProcessGroupGloo.cpp.o [1123/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/ProcessGroupMPI.cpp.o [1124/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/ProcessGroupWrapper.cpp.o [1125/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/Store.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1126/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/ProcessGroup.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1127/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/TCPStore.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1128/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/TCPStoreLibUvBackend.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreLibUvBackend.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1129/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/TCPStoreBackend.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStore.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreBackend.hpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/TCPStoreBackend.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1130/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/Utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.hpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1131/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/debug.cpp.o [1132/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/comm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.hpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1133/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/default_comm_hooks.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/default_comm_hooks.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/default_comm_hooks.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1134/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/logger.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/logger.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Utils.hpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1135/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/logging.cpp.o [1136/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/quantization/quantization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/quantization/quantization.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1137/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/sequence_num.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/sequence_num.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1138/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/socket.cpp.o [1139/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/reducer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.hpp:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer.hpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1140/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/control_plane/Handlers.cpp.o [1141/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/Work.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Work.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1142/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/control_plane/WorkerServer.cpp.o [1143/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/autograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1144/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/utils.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1145/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/context/container.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1146/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/context/context.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1147/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/engine/dist_engine.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CachedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/engine/dist_engine.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/engine/dist_engine.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1148/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/functions/sendrpc_backward.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/sendrpc_backward.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/sendrpc_backward.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1149/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/autograd_metadata.cpp.o [1150/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/functions/recvrpc_backward.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1151/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1152/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1153/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1154/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1155/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1156/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1157/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1158/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1159/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/HashStore.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/HashStore.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/HashStore.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1160/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1161/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.hpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1162/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/agent_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1163/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/message.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1164/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/server_process_global_profiler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1165/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1166/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/python_call.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_call.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_call.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1167/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/python_remote_call.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_remote_call.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1168/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/python_resp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_resp.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_resp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1169/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/request_callback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1170/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/request_callback_no_python.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_no_python.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1171/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/rpc_agent.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1172/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/rref_context.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1173/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/rref_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1174/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/rref_proto.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_proto.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_proto.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1175/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/script_call.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_call.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_call.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1176/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/script_remote_call.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_remote_call.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_remote_call.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1177/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/script_resp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/script_resp.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1178/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/tensorpipe_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1179/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/tensorpipe_agent.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/agent_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp: In function ‘void torch::distributed::rpc::{anonymous}::makeStreamsWaitOnOthers(const std::vector&, const std::vector&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp:146:24: warning: possibly dangling reference to a temporary [-Wdangling-reference] 146 | const c10::Stream& consumer = | ^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.cpp:147:27: note: the temporary was destroyed at the end of the full expression ‘torch::distributed::rpc::getStreamForDevice((* & consumers), (& producer)->c10::Stream::device())’ 147 | getStreamForDevice(consumers, producer.device()); | ~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [1180/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/faulty_tensorpipe_agent.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1181/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/types.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1182/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/torchscript_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/record_function_ops.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1183/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/cuda.cpp.o [1184/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/data/datasets/mnist.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/example.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/datasets/mnist.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/datasets/mnist.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1185/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/distributed/rpc/utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1186/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/data/samplers/distributed.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/distributed.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/distributed.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1187/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/data/samplers/random.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/random.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/random.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1188/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/data/samplers/sequential.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/sequential.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/sequential.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1189/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/enum.cpp.o [1190/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/data/samplers/stream.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/base.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/data/samplers/stream.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/data/samplers/stream.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1191/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/imethod.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/imethod.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/imethod.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1192/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/mps.cpp.o [1193/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/init.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/init.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1194/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/serialize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/sugared_value.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/resolver.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1195/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/module.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1196/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/_functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/_functions.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/_functions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/normalization.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/_functions.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1197/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/adaptive.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/adaptive.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/adaptive.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1198/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/activation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/dropout.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/dropout.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/activation.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/activation.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1199/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/batchnorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/batchnorm.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/batchnorm.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/batchnorm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1200/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/normalization.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/normalization.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1201/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/instancenorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/batchnorm.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/instancenorm.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/instancenorm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/instancenorm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1202/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/conv.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/conv.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/conv.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/conv.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1203/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/dropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/dropout.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/dropout.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/dropout.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1204/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/distance.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/distance.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/distance.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1205/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/embedding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/embedding.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/embedding.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1206/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/fold.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/fold.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/fold.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1207/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/functional/linear.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/linear.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1208/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/loss.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/loss.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/loss.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1209/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/padding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/padding.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/padding.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1210/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/pixelshuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/pixelshuffle.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/pixelshuffle.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1211/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/pooling.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/pooling.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1212/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/rnn.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/rnn.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/rnn.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1213/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/upsampling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/upsampling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/upsampling.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1214/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/transformer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/init.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/transformer.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1215/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/modules/container/functional.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/functional.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/modules/container/functional.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/cloneable.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/functional.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1216/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/activation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/activation.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/activation.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1217/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/adaptive.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/adaptive.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/adaptive.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1218/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/batchnorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/batchnorm.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/batchnorm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1219/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/embedding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/embedding.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/embedding.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1220/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/instancenorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/batchnorm.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/instancenorm.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/instancenorm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1221/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/normalization.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/normalization.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1222/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/dropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/dropout.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/dropout.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1223/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/conv.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/conv.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/conv.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1224/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/padding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/padding.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/padding.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1225/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/linear.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/linear.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1226/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/pooling.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/pooling.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1227/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/rnn.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/rnn.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/rnn.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1228/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/vision.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/vision.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/vision.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1229/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/nn/options/transformer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/options/transformer.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/nn/options/transformer.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1230/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/adagrad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/pimpl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adagrad.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adagrad.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1231/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/adam.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adam.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adam.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1232/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/adamw.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/adamw.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/adamw.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1233/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/lbfgs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/lbfgs.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/lbfgs.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1234/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/optimizer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/TensorDataContainer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/variable_factories.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/optimizer.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/optimizer.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1235/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/rmsprop.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/rmsprop.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/rmsprop.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1236/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/serialize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/input-archive.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/archive.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/serialize.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/serialize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1237/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/schedulers/lr_scheduler.cpp.o [1238/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/schedulers/step_lr.cpp.o [1239/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/schedulers/reduce_on_plateau_scheduler.cpp.o [1240/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/optim/sgd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/nn/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/optim/sgd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/optim/sgd.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1241/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/serialize/input-archive.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/input-archive.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize/input-archive.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1242/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/xpu.cpp.o [1243/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/api/src/serialize/output-archive.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/serialize/output-archive.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/serialize/output-archive.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1244/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/UfuncCPUKernel_add.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/AccumulateType.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Math.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec_base.h:27, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional_base.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ufunc/add.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/UfuncCPUKernel_add.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/UfuncCPUKernel_add.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1245/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/spherical_bessel_j0.cpp.AVX2.cpp.o [1246/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k1.cpp.AVX2.cpp.o [1247/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k0.cpp.AVX2.cpp.o [1248/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1249/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int8mm_kernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int8mm_kernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int8mm_kernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1250/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int4mm_kernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int4mm_kernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int4mm_kernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1251/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1252/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/batch_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1253/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/group_norm_kernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/group_norm_kernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ In file included from /usr/include/string.h:548, from /usr/include/c++/14/cstring:43, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/string_view.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/StringUtil.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/Exception.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Device.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:11: In function ‘void* memcpy(void*, const void*, size_t)’, inlined from ‘void at::vec::AVX2::Vectorized::store(void*, int) const’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h:91:18, inlined from ‘void at::vec::AVX2::Vectorized::store(void*, int) const’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h:85:8, inlined from ‘void at::vec::map2(const Op&, scalar_t*, const scalar_t*, const scalar_t*, int64_t) [with scalar_t = double; Op = at::native::{anonymous}::CalcMeanVar(const double*, double*, double*, int64_t)::; typename std::enable_if<(! is_reduced_floating_point_v), int>::type = 0]’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional_base.h:297:21, inlined from ‘typename std::enable_if, >::value>::type at::native::{anonymous}::CalcMeanVar(const T*, opmath_t*, opmath_t*, int64_t) [with T = double; opmath_t = double]’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:170:15, inlined from ‘at::native::{anonymous}::GroupNormKernelImplChannelsLastInternal(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&)::’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:413:33, inlined from ‘at::parallel_for(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&):: >(int64_t, int64_t, int64_t, const native::{anonymous}::GroupNormKernelImplChannelsLastInternal(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&)::&)::’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Parallel-inl.h:36:10, inlined from ‘_ZN2at8internal15invoke_parallelIZNS_12parallel_forIZNS_6native12_GLOBAL__N_139GroupNormKernelImplChannelsLastInternalIddEEvRKNS_6TensorES8_S8_lllldRS6_S9_S9_EUlllE0_EEvlllRKT_EUlllE_EEvlllSD_._omp_fn.0’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelOpenMP.h:41:10: /usr/include/bits/string_fortified.h:29:33: warning: ‘void* __builtin_memcpy(void*, const void*, long unsigned int)’ reading between 40 and 17179869176 bytes from a region of size 32 [-Wstringop-overread] 29 | return __builtin___memcpy_chk (__dest, __src, __len, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~ 30 | __glibc_objsize0 (__dest)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h: In function ‘_ZN2at8internal15invoke_parallelIZNS_12parallel_forIZNS_6native12_GLOBAL__N_139GroupNormKernelImplChannelsLastInternalIddEEvRKNS_6TensorES8_S8_lllldRS6_S9_S9_EUlllE0_EEvlllRKT_EUlllE_EEvlllSD_._omp_fn.0’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h:89:14: note: source object ‘tmp_values’ of size 32 89 | double tmp_values[size()]; | ^~~~~~~~~~ In function ‘void* memcpy(void*, const void*, size_t)’, inlined from ‘void at::vec::AVX2::Vectorized::store(void*, int) const’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h:91:18, inlined from ‘void at::vec::AVX2::Vectorized::store(void*, int) const’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h:85:8, inlined from ‘void at::vec::map2(const Op&, scalar_t*, const scalar_t*, const scalar_t*, int64_t) [with scalar_t = double; Op = at::native::{anonymous}::CalcMeanVar(const double*, double*, double*, int64_t)::; typename std::enable_if<(! is_reduced_floating_point_v), int>::type = 0]’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional_base.h:297:21, inlined from ‘typename std::enable_if, >::value>::type at::native::{anonymous}::CalcMeanVar(const T*, opmath_t*, opmath_t*, int64_t) [with T = double; opmath_t = double]’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:170:15, inlined from ‘at::native::{anonymous}::GroupNormKernelImplChannelsLastInternal(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&)::’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:413:33, inlined from ‘void at::parallel_for(int64_t, int64_t, int64_t, const F&) [with F = native::{anonymous}::GroupNormKernelImplChannelsLastInternal(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&)::]’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Parallel-inl.h:29:6, inlined from ‘void at::native::{anonymous}::GroupNormKernelImplChannelsLastInternal(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&) [with T = double; PT = double]’ at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:403:21: /usr/include/bits/string_fortified.h:29:33: warning: ‘void* __builtin_memcpy(void*, const void*, long unsigned int)’ reading between 40 and 17179869176 bytes from a region of size 32 [-Wstringop-overread] 29 | return __builtin___memcpy_chk (__dest, __src, __len, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~ 30 | __glibc_objsize0 (__dest)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h: In function ‘void at::native::{anonymous}::GroupNormKernelImplChannelsLastInternal(const at::Tensor&, const at::Tensor&, const at::Tensor&, int64_t, int64_t, int64_t, int64_t, double, at::Tensor&, at::Tensor&, at::Tensor&) [with T = double; PT = double]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256_double.h:89:14: note: source object ‘tmp_values’ of size 32 89 | double tmp_values[size()]; | ^~~~~~~~~~ [1254/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/airy_ai.cpp.AVX2.cpp.o [1255/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/WeightNormKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/WeightNormKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/WeightNormKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1256/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1257/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1258/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernelAVXAntialias.h:32, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1259/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Unfold2d.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/Unfold2d.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/Unfold2d.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1260/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1261/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SumKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/AccumulateType.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SumKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SumKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1262/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/StackKernel.cpp.AVX2.cpp.o [1263/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Generator.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnaryOps.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1264/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1265/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SparseFactories.cpp.AVX2.cpp.o [1266/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SoftMaxKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1267/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SortingKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SortingKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SortingKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1268/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NonEmptyUtils.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1269/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RenormKernel.cpp.AVX2.cpp.o [1270/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1271/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOpsUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1272/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1273/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1274/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PowKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PowKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PowKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1275/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PixelShuffleKernel.cpp.AVX2.cpp.o [1276/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1277/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp.AVX2.cpp.o [1278/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MultinomialKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MultinomialKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1279/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxUnpoolKernel.cpp.AVX2.cpp.o [1280/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPooling.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MaxPooling.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MaxPooling.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1281/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PaddingKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PaddingKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PaddingKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1282/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptivePooling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MaxPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1283/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1284/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LerpKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/OpMathType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Lerp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/LerpKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/LerpKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1285/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/HistogramKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/aminmax.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/HistogramKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Histogram.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1286/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/IndexKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/IndexKernel.cpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/IndexKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1287/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedSGDKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1288/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedAdamKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1289/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1290/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1291/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/GridSamplerKernel.cpp.AVX2.cpp.o [1292/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FillKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch_v2.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FillKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FillKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1293/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistributionKernels.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DistributionKernels.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1294/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistanceOpsKernel.cpp.AVX2.cpp.o [1295/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1296/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CrossKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CrossKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CrossKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1297/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1298/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ComplexKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ComplexKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ComplexKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1299/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ChannelShuffleKernel.cpp.AVX2.cpp.o [1300/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CatKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CatKernel.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CatKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CatKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1301/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/BlasKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/BlasKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/BlasKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1302/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CopyKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CopyKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CopyKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1303/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AvgPoolKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AvgPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AvgPoolKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1304/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1305/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1306/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BinaryOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1307/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1308/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/UfuncCPUKernel_add.cpp.AVX512.cpp.o [1309/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1310/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/spherical_bessel_j0.cpp.AVX512.cpp.o [1311/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k1.cpp.AVX512.cpp.o [1312/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k0.cpp.AVX512.cpp.o [1313/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1314/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Activation.cpp.AVX2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/Activation.cpp:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/Activation.cpp.AVX2.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1315/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int8mm_kernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int8mm_kernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int8mm_kernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1316/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/group_norm_kernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/group_norm_kernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1317/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int4mm_kernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int4mm_kernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int4mm_kernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1318/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/airy_ai.cpp.AVX512.cpp.o [1319/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1320/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/WeightNormKernel.cpp.AVX512.cpp.o [1321/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.AVX512.cpp.o [1322/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1323/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1324/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Unfold2d.cpp.AVX512.cpp.o [1325/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1326/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SumKernel.cpp.AVX512.cpp.o [1327/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/StackKernel.cpp.AVX512.cpp.o [1328/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1329/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SparseFactories.cpp.AVX512.cpp.o [1330/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SortingKernel.cpp.AVX512.cpp.o [1331/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Generator.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnaryOps.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1332/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1333/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SoftMaxKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1334/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RenormKernel.cpp.AVX512.cpp.o [1335/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1336/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1337/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOpsUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1338/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.AVX512.cpp.o [1339/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.AVX512.cpp.o [1340/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PixelShuffleKernel.cpp.AVX512.cpp.o [1341/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PowKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PowKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PowKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1342/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PaddingKernel.cpp.AVX512.cpp.o [1343/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp.AVX512.cpp.o [1344/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MultinomialKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MultinomialKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1345/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxUnpoolKernel.cpp.AVX512.cpp.o [1346/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPooling.cpp.AVX512.cpp.o [1347/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.AVX512.cpp.o [1348/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.AVX512.cpp.o [1349/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LerpKernel.cpp.AVX512.cpp.o [1350/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/IndexKernel.cpp.AVX512.cpp.o [1351/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/HistogramKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/aminmax.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/HistogramKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1352/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/GridSamplerKernel.cpp.AVX512.cpp.o [1353/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.AVX512.cpp.o [1354/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.AVX512.cpp.o [1355/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.AVX512.cpp.o [1356/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.AVX512.cpp.o [1357/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FillKernel.cpp.AVX512.cpp.o [1358/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistributionKernels.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DistributionKernels.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1359/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistanceOpsKernel.cpp.AVX512.cpp.o [1360/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1361/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CrossKernel.cpp.AVX512.cpp.o [1362/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FlashAttentionKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1363/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ComplexKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ComplexKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ComplexKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1364/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ChannelShuffleKernel.cpp.AVX512.cpp.o [1365/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CatKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CatKernel.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CatKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CatKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1366/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/BlasKernel.cpp.AVX512.cpp.o [1367/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BinaryOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/BinaryOpsKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1368/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AvgPoolKernel.cpp.AVX512.cpp.o [1369/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/AmpGradScalerKernels.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1370/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CopyKernel.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CopyKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CopyKernel.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1371/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AdaptiveMaxPoolKernel.cpp.AVX512.cpp.o [1372/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/AdaptiveAvgPoolKernel.cpp.AVX512.cpp.o [1373/1912] Linking CXX static library lib/libCaffe2_perfkernels_avx.a [1374/1912] Linking CXX static library lib/libCaffe2_perfkernels_avx2.a [1375/1912] Linking CXX static library lib/libCaffe2_perfkernels_avx512.a [1376/1912] Linking CXX shared library lib/libc10_hip.so Warning: Unused direct dependencies: /lib64/libm.so.6 [1377/1912] Generating /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/__init__.pyi, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/_C/_VariableFunctions.pyi, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/nn/functional.pyi [1378/1912] Generating /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/utils/data/datapipes/datapipe.pyi [1379/1912] Generating /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/version.py [1380/1912] Building CXX object caffe2/torch/lib/libshm/CMakeFiles/shm.dir/core.cpp.o [1381/1912] Building CXX object caffe2/CMakeFiles/caffe2_nvrtc.dir/__/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /usr/include/hip/hip_runtime.h:70, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/ATenHIPGeneral.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:4: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp: In function ‘at::cuda::NVRTC* at::cuda::load_nvrtc()’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp:9:42: note: in definition of macro ‘CREATE_ASSIGN’ 9 | #define CREATE_ASSIGN(name) self->name = name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp:10:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 10 | AT_FORALL_NVRTC(CREATE_ASSIGN) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp:9:42: note: in definition of macro ‘CREATE_ASSIGN’ 9 | #define CREATE_ASSIGN(name) self->name = name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.cpp:10:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 10 | AT_FORALL_NVRTC(CREATE_ASSIGN) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1382/1912] Building C object caffe2/CMakeFiles/torch_global_deps.dir/__/torch/csrc/empty.c.o [1383/1912] Building CXX object caffe2/torch/lib/libshm/CMakeFiles/torch_shm_manager.dir/manager.cpp.o [1384/1912] Linking CXX shared library lib/libcaffe2_nvrtc.so Warning: Unused direct dependencies: /lib64/libhiprtc.so.6 /lib64/libm.so.6 /lib64/libgcc_s.so.1 [1385/1912] Linking C shared library lib/libtorch_global_deps.so [1386/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Activation.cpp.AVX512.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/Activation.cpp:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/Activation.cpp.AVX512.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1387/1912] Linking CXX shared library lib/libtorch_cpu.so [1388/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/torch_hip_generated_Sleep.hip.o [1389/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/build/aten/src/ATen/torch_hip_generated_UfuncCUDA_add.cu.o [1390/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/torch_hip_generated_cub-RadixSortKeys.hip.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1010. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1012. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1030. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1031. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1100. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1101. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for gfx1102. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:54:41: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 54 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTATIATE_CUB_TEMPLATES) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:55:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 55 | AT_INSTATIATE_CUB_TEMPLATES(uint16_t, UInt16) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:56:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 56 | AT_INSTATIATE_CUB_TEMPLATES(uint32_t, UInt32) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:57:1: note: in instantiation of function template specialization 'at::cuda::cub::radix_sort_keys' requested here 57 | AT_INSTATIATE_CUB_TEMPLATES(uint64_t, UInt64) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:46:17: note: expanded from macro 'AT_INSTATIATE_CUB_TEMPLATES' 46 | template void radix_sort_keys( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:26:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 25 | CUB_WRAPPER( | ~~~~~~~~~~~~ 26 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeysDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 27 | keys_in_, | ~~~~~~~~~ 28 | keys_out_, | ~~~~~~~~~~ 29 | n, | ~~ 30 | begin_bit, | ~~~~~~~~~~ 31 | end_bit, | ~~~~~~~~ 32 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortKeys.hip:35:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 34 | CUB_WRAPPER( | ~~~~~~~~~~~~ 35 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortKeys, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 36 | keys_in_, | ~~~~~~~~~ 37 | keys_out_, | ~~~~~~~~~~ 38 | n, | ~~ 39 | begin_bit, | ~~~~~~~~~~ 40 | end_bit, | ~~~~~~~~ 41 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 warnings generated when compiling for host. [1391/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/torch_hip_generated_cub.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:20:3: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 20 | inclusive_scan(input, output, Sum{}, num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, int, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:30:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long, 1073741824>' requested here 30 | exclusive_scan(input, output, SumOp{}, scalar_t(0), num_items); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:48:3: note: in instantiation of function template specialization 'at::cuda::cub::exclusive_scan, long *, at::cuda::cub::(anonymous namespace)::SumOp, long, 1073741824>' requested here 48 | exclusive_scan(iter, output_idx, SumOp{}, int64_t{0}, n); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.hip:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:307:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 307 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::ExclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 308 | input, | ~~~~~~ 309 | output, | ~~~~~~~ 310 | scan_op, | ~~~~~~~~ 311 | init_value, | ~~~~~~~~~~~ 312 | num_items, | ~~~~~~~~~~ 313 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for host. [1392/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/detail/torch_hip_generated_IndexUtils.hip.o [1393/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/torch_hip_generated_jiterator.hip.o [1394/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AbsKernel.hip.o [1395/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/torch_hip_generated_cub-RadixSortPairs.hip.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1010. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1012. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1030. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1031. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1100. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1101. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for gfx1102. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:70:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 70 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:71:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 71 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:72:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 72 | AT_INSTANTIATE_SORT_PAIRS(int32_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:73:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 73 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 1) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:74:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 74 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 2) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:75:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 75 | AT_INSTANTIATE_SORT_PAIRS(int64_t, 4) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:80:41: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 80 | AT_FORALL_SCALAR_TYPES_AND2(Bool, Half, AT_INSTANTIATE_SORT_PAIRS_8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:81:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 81 | AT_INSTANTIATE_SORT_PAIRS(uint16_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:82:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 82 | AT_INSTANTIATE_SORT_PAIRS(uint32_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:83:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 83 | AT_INSTANTIATE_SORT_PAIRS(uint64_t, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:84:1: note: in instantiation of function template specialization 'at::cuda::cub::detail::radix_sort_pairs_impl' requested here 84 | AT_INSTANTIATE_SORT_PAIRS(c10::BFloat16, 8) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:60:17: note: expanded from macro 'AT_INSTANTIATE_SORT_PAIRS' 60 | template void radix_sort_pairs_impl( \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:36:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 35 | CUB_WRAPPER( | ~~~~~~~~~~~~ 36 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 37 | keys_in_, | ~~~~~~~~~ 38 | keys_out_, | ~~~~~~~~~~ 39 | values_in, | ~~~~~~~~~~ 40 | values_out, | ~~~~~~~~~~~ 41 | n, | ~~ 42 | begin_bit, | ~~~~~~~~~~ 43 | end_bit, | ~~~~~~~~ 44 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub-RadixSortPairs.hip:47:32: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 46 | CUB_WRAPPER( | ~~~~~~~~~~~~ 47 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 48 | keys_in_, | ~~~~~~~~~ 49 | keys_out_, | ~~~~~~~~~~ 50 | values_in, | ~~~~~~~~~~ 51 | values_out, | ~~~~~~~~~~~ 52 | n, | ~~ 53 | begin_bit, | ~~~~~~~~~~ 54 | end_bit, | ~~~~~~~~ 55 | c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 76 warnings generated when compiling for host. [1396/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationEluKernel.hip.o [1397/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationGeluKernel.hip.o [1398/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationGluKernel.hip.o [1399/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationHardshrinkKernel.hip.o [1400/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationHardsigmoidKernel.hip.o [1401/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationHardswishKernel.hip.o [1402/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationHardtanhKernel.hip.o [1403/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationLeakyReluKernel.hip.o [1404/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationLogSigmoidKernel.hip.o [1405/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationPreluKernel.hip.o [1406/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationMishKernel.hip.o [1407/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationSiluKernel.hip.o [1408/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationSoftplusKernel.hip.o [1409/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationSoftshrinkKernel.hip.o [1410/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AdaptiveAveragePooling.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1411/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ActivationThresholdKernel.hip.o [1412/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AdaptiveAveragePooling3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveAveragePooling3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1413/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AdaptiveMaxPooling2d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling2d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1414/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AdaptiveMaxPooling3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AdaptiveMaxPooling3d.hip:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1415/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AmpKernels.hip.o [1416/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AveragePool2d.hip.o [1417/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_AveragePool3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/AveragePool3d.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1418/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryBitwiseOpsKernels.hip.o [1419/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryDivFloorKernel.hip.o [1420/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryDivTrueKernel.hip.o [1421/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryDivTruncKernel.hip.o [1422/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryGeometricKernels.hip.o [1423/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryMiscBackwardOpsKernels.hip.o [1424/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryLogicalOpsKernels.hip.o [1425/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryMiscOpsKernels.hip.o [1426/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryMulKernel.hip.o [1427/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryShiftOpsKernels.hip.o [1428/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Bucketization.hip.o [1429/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_BinaryRemainderKernel.hip.o [1430/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Col2Im.hip.o [1431/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CompareKernels.hip.o [1432/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ComplexKernel.hip.o [1433/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CompareEQKernel.hip.o [1434/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ConvolutionMM2d.hip.o [1435/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CopysignKernel.hip.o [1436/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CrossKernel.hip.o [1437/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Copy.hip.o [1438/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CumminmaxKernel.hip.o [1439/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CumprodKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::multiplies>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::multiplies>>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:15:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 15 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumprodKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for host. [1440/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_CumsumKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, std::plus>, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim, std::plus>>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan, 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:17:9: note: in instantiation of function template specialization 'at::native::scan_dim>' requested here 17 | scan_dim( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/CumsumKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 22 warnings generated when compiling for host. [1441/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DepthwiseConv2d.hip.o [1442/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DepthwiseConv3d.hip.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1010. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1012. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1030. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1031. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1100. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1101. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 184 | conv_depthwise3d_cuda_backward_weight_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DepthwiseConv3d.hip:184:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 12 warnings generated when compiling for gfx1102. [1443/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DilatedMaxPool2d.hip.o [1444/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DilatedMaxPool3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DilatedMaxPool3d.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1445/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistanceKernel.hip.o [1446/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionCauchyKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionCauchyKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1447/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionBernoulli.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionBernoulli.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1448/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionExponentialKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionExponentialKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1449/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionGeometricKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionGeometricKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1450/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionLogNormalKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionLogNormalKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1451/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionNormal.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionNormal.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1452/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionUniform.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionUniform.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1453/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_DistributionRandomKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionRandomKernel.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1454/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Dropout.hip.o [1455/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Embedding.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:365:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 365 | cuda::cub::unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Embedding.hip:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 5 warnings generated when compiling for host. [1456/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Distributions.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.hip:5: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1457/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_EmbeddingBackwardKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBackwardKernel.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1458/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_EmbeddingBag.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/EmbeddingBag.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1459/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FillKernel.hip.o [1460/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FlattenIndicesKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FlattenIndicesKernel.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1461/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachBinaryOpScalar.hip.o [1462/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachBinaryOpList.hip.o [1463/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachBinaryOpScalarList.hip.o [1464/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachBinaryOpScalarTensor.hip.o [1465/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachPointwiseOp.hip.o [1466/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachReduceOp.hip.o [1467/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachTernaryOp.hip.o [1468/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FractionalMaxPool2d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool2d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1469/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FractionalMaxPool3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FractionalMaxPool3d.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1470/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FunctionOfAMatrixUtilsKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/FunctionOfAMatrixUtilsKernel.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1471/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FusedAdamKernel.hip.o [1472/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FusedAdamWKernel.hip.o [1473/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_FusedSgdKernel.hip.o [1474/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_GcdLcmKernel.hip.o [1475/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ForeachUnaryOp.hip.o [1476/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_HIPScalar.hip.o [1477/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_IGammaKernel.hip.o [1478/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Im2Col.hip.o [1479/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_GridSampler.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:3: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:181:11: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 181 | coord = compute_coordinates(coord, size, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:66:21: note: in instantiation of function template specialization 'at::native::grid_sampler_compute_source_index' requested here 66 | opmath_t ix = grid_sampler_compute_source_index(x, inp_W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:144:18: warning: implicit conversion from 'int' to 'float' changes value from 2147483646 to 2147483648 [-Wimplicit-const-int-float-conversion] 144 | if (x > INT_MAX-1 || x < INT_MIN || !::isfinite(static_cast(x))) | ~ ~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:168:11: note: in instantiation of function template specialization 'at::native::safe_downgrade_to_int_range' requested here 168 | coord = safe_downgrade_to_int_range(coord); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cuh:236:7: note: in instantiation of function template specialization 'at::native::compute_coordinates' requested here 236 | x = compute_coordinates(x, W, padding_mode, align_corners); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:138:15: note: in instantiation of function template specialization 'at::native::get_value_bounded' requested here 138 | get_value_bounded(inp_ptr_NC, ix_nw - 1, iy_nw - 1 + i, inp_W, inp_H, inp_sW, inp_sH, padding_mode, align_corners), | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.hip:770:29: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::grid_sampler_2d_kernel' requested here 770 | hipLaunchKernelGGL(( grid_sampler_2d_kernel) | ^ 4 warnings generated when compiling for host. [1480/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_IndexKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.hip:16: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1481/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_LegacyThrustHelpers.hip.o [1482/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Lerp.hip.o [1483/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Indexing.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Indexing.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1484/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_LinearAlgebra.hip.o [1485/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_LogAddExpKernel.hip.o [1486/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_LogcumsumexpKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan *, c10::complex *, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3), 1073741824>' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim, (lambda at /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:112:3)>' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:453:16: note: in instantiation of function template specialization 'at::cuda::cub::inclusive_scan' requested here 453 | cuda::cub::inclusive_scan(self_->const_data_ptr(), result.mutable_data_ptr(), binary_op, self.numel()); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:121:9: note: in instantiation of function template specialization 'at::native::scan_dim' requested here 121 | scan_dim(self, result, dim, init, log_add_exp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:110:24: note: expanded from macro '_LCME_DISPATCH' 110 | #define _LCME_DISPATCH AT_DISPATCH_FLOATING_AND_COMPLEX_TYPES_AND2 | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LogcumsumexpKernel.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanUtils.cuh:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:240:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 240 | CUB_WRAPPER(NO_ROCM(detail)::hipcub::DeviceScan::InclusiveScan, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | input, | ~~~~~~ 242 | output, | ~~~~~~~ 243 | scan_op, | ~~~~~~~~ 244 | num_items, | ~~~~~~~~~~ 245 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 12 warnings generated when compiling for host. [1487/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Loss.hip.o [1488/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_LossCTC.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LossCTC.hip:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1489/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_MaxUnpooling.hip.o [1490/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_MixedDtypesLinear.hip.o [1491/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_MultiLabelMarginCriterion.hip.o [1492/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_MultiMarginLoss.hip.o [1493/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_MaxMinElementwiseKernel.hip.o [1494/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_MultinomialKernel.hip.o [1495/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_NLLLoss2d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/NLLLoss2d.hip:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1496/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_NaiveConvolutionTranspose2d.hip.o [1497/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_NaiveConvolutionTranspose3d.hip.o [1498/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_NaiveDilatedConvolution.hip.o [1499/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Nonzero.hip.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1010. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1012. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1030. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1031. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1100. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1101. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for gfx1102. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl>' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:71:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 71 | hipcub::DeviceReduce::Sum(nullptr, temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:124:10: note: in instantiation of function template specialization 'at::native::nonzero_cuda_out_impl' requested here 124 | [&] {nonzero_cuda_out_impl(self, out);}); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:73:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 73 | hipcub::DeviceReduce::Sum(temp_storage.get(), temp_storage_bytes, itr, (int*)num_nonzeros.get(), N, stream); | ^~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:89:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 89 | hipcub::DeviceSelect::Flagged(nullptr, temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 90 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Nonzero.hip:92:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 92 | hipcub::DeviceSelect::Flagged(temp_storage.get(), temp_storage_bytes, counting_itr, itr, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 93 | out_temp.mutable_data_ptr(), (int*)num_nonzeros.get(), N, stream); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 52 warnings generated when compiling for host. [1500/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Normalization.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.hip:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 1169 | __global__ void batch_norm_backward_reduce_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:1169:17: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 946 | batch_norm_collect_statistics_channels_last_kernel( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Normalization.cuh:946:1: warning: loop not unrolled: the optimizer was unable to perform the requested transformation; the transformation might be disabled or specified as part of an unsupported transformation ordering [-Wpass-failed=transform-warning] 28 warnings generated when compiling for gfx1102. [1501/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_PointwiseOpsKernel.hip.o [1502/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_RNN.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:9: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RNN.hip:61:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 61 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ 2 warnings generated when compiling for host. [1503/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Randperm.hip.o [1504/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_RangeFactories.hip.o [1505/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_RecordStream.hip.o [1506/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_PowKernel.hip.o [1507/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Reduce.hip.o [1508/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceAMinMaxKernel.hip.o [1509/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceArgMaxKernel.hip.o [1510/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceArgMinKernel.hip.o [1511/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceLogicKernel.hip.o [1512/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceMaxValuesKernel.hip.o [1513/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceMinValuesKernel.hip.o [1514/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceMomentKernel.hip.o [1515/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceSumProdKernel.hip.o [1516/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReflectionPad.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReflectionPad.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1517/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_RenormKernel.hip.o [1518/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Repeat.hip.o [1519/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReplicationPadding.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReplicationPadding.hip:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1520/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_RreluWithNoise.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/RreluWithNoise.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/DistributionTemplates.h:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1521/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ScatterGatherKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScatterGatherKernel.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1522/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SegmentReduce.hip.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1010. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1012. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1030. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1031. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1100. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1101. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for gfx1102. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:795:41: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:267:35: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 267 | AT_DISPATCH_CASE_FLOATING_TYPES(__VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 4 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:268:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 268 | AT_DISPATCH_CASE(SCALARTYPE1, __VA_ARGS__) \ | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:498:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_stride_axis | ~~~~~~~~~~~~~~~~~~~ 442 | ); | ~~ 443 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 444 | } else { | ~~~~~~~~ 445 | if (reduction == ReductionType::MAX) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMax max_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | max_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:510:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | max_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MEAN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:534:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | segment_count, | ~~~~~~~~~~~~~~ 442 | initial.has_value(), | ~~~~~~~~~~~~~~~~~~~~ 443 | initial_value); | ~~~~~~~~~~~~~~~ 444 | C10_HIP_KERNEL_LAUNCH_CHECK(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::MIN) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomMin min_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | min_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:546:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | min_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::SUM) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomSum sum_op{}; | ~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | sum_op, | ~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SegmentReduce.hip:558:23: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 441 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 442 | sum_op, | ~~~~~~~ 443 | initial_value, | ~~~~~~~~~~~~~~ 444 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 445 | } else if (reduction == ReductionType::PROD) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 446 | CustomProd prod_op{}; | ~~~~~~~~~~~~~~~~~~~~~ 447 | CUB_WRAPPER( | ~~~~~~~~~~~~ 448 | hipcub::DeviceSegmentedReduce::Reduce, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 449 | data_data_ptr, | ~~~~~~~~~~~~~~ 450 | output_data_ptr, | ~~~~~~~~~~~~~~~~ 451 | segment_count, | ~~~~~~~~~~~~~~ 452 | offsets_data_ptr, | ~~~~~~~~~~~~~~~~~ 453 | offsets_data_ptr + 1, | ~~~~~~~~~~~~~~~~~~~~~ 454 | prod_op, | ~~~~~~~~ 455 | initial_value, | ~~~~~~~~~~~~~~ 456 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:277:37: note: expanded from macro 'AT_DISPATCH_FLOATING_TYPES_AND2' 277 | SCALARTYPE1, SCALARTYPE2, __VA_ARGS__)) | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:269:33: note: expanded from macro 'AT_DISPATCH_CASE_FLOATING_TYPES_AND2' 269 | AT_DISPATCH_CASE(SCALARTYPE2, __VA_ARGS__) | ^~~~~~~~~~~ note: (skipping 3 expansions in backtrace; use -fmacro-backtrace-limit=0 to see all) /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:797:46: note: expanded from macro 'AT_DISPATCH_INDEX_TYPES' 791 | AT_DISPATCH_SWITCH( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 792 | TYPE, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 793 | NAME, \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 794 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 795 | at::ScalarType::Int, index_t, __VA_ARGS__) \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 796 | AT_PRIVATE_CASE_TYPE_USING_HINT( \ | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 797 | at::ScalarType::Long, index_t, __VA_ARGS__)) | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:70:12: note: expanded from macro 'AT_PRIVATE_CASE_TYPE_USING_HINT' 70 | return __VA_ARGS__(); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:221:7: note: expanded from macro 'AT_DISPATCH_SWITCH' 221 | __VA_ARGS__ \ | ^~~~~~~~~~~ 80 warnings generated when compiling for host. [1523/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Shape.hip.o [1524/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ReduceNormKernel.hip.o [1525/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SoftMax.hip.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1010. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1012. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1030. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1031. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1100. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1101. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for gfx1102. /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:934:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 934 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:943:17: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 943 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1074:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax' requested here 1074 | host_softmax(input, dim, half_to_float, output); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1040:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1040 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:152:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 152 | hipOccupancyMaxActiveBlocksPerMultiprocessor(&max_active_blocks, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~ 153 | k, block_threads, smem_size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1052:11: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::SpatialSoftMax_getLaunchSizes' requested here 1052 | SpatialSoftMax_getLaunchSizes( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SoftMax.hip:1090:3: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::host_softmax_backward' requested here 1090 | host_softmax_backward(grad, output, dim, half_to_float, grad_input); | ^ 18 warnings generated when compiling for host. [1526/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SortImpl.hip.o [1527/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Sort.hip.o [1528/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Sorting.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1529/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SortStable.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:206:18: note: in instantiation of function template specialization 'at::cuda::cub::segmented_sort_pairs' requested here 206 | at::cuda::cub::segmented_sort_pairs( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:268:13: note: in instantiation of function template specialization 'at::native::(anonymous namespace)::segmented_sort_pairs' requested here 268 | segmented_sort_pairs( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortStable.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:151:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 151 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairsDescending, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 152 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 153 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 154 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:156:40: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 156 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSegmentedRadixSort::SortPairs, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 157 | keys_in_, keys_out_, values_in, values_out, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 158 | num_elements, num_segments, begin_offsets, end_offsets, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 159 | begin_bit, end_bit, c10::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 40 warnings generated when compiling for host. [1530/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SparseMM.hip.o [1531/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SpectralOps.hip.o [1532/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SparseBinaryOpIntersectionKernel.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SparseBinaryOpIntersectionKernel.hip:7: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/KernelUtils.cuh:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1533/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_SummaryOps.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SummaryOps.hip:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1534/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_StepKernel.hip.o [1535/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TensorFactories.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorFactories.hip:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1536/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TensorCompare.hip.o [1537/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TensorShape.hip.o [1538/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TensorTopK.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.hip:15: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SortingRadixSelect.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1539/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TensorModeKernel.hip.o [1540/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TensorTransformations.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTransformations.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1541/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryComplexKernels.hip.o [1542/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_TriangularOps.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TriangularOps.hip:22: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1543/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGammaKernels.hip.o [1544/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricAcosKernel.hip.o [1545/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryFractionKernels.hip.o [1546/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricAsinKernel.hip.o [1547/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricAcoshKernel.hip.o [1548/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricAtanKernel.hip.o [1549/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricAsinhKernel.hip.o [1550/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricAtanhKernel.hip.o [1551/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricCosKernel.hip.o [1552/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricCoshKernel.hip.o [1553/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricSinKernel.hip.o [1554/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricSinhKernel.hip.o [1555/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricTanKernel.hip.o [1556/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryGeometricTanhKernel.hip.o [1557/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryLogKernels.hip.o [1558/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnaryOpsKernel.hip.o [1559/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnarySignKernels.hip.o [1560/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnfoldBackwardKernel.hip.o [1561/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UnarySpecialOpsKernel.hip.o [1562/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UniqueCub.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:8: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique, bool *, long *>' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode, bool *, long *, long *>' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:252:14: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 252 | return compute_unique( | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:266:20: note: in instantiation of function template specialization 'at::cuda::cub::reduce, int *, hipcub::Sum, int>' requested here 266 | at::cuda::cub::reduce(data_iter, tmp_num_true.get(), num_inp, | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:408:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 407 | CUB_WRAPPER( | ~~~~~~~~~~~~ 408 | NO_ROCM(at_cuda_detail)::hipcub::DeviceReduce::Reduce, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 409 | input, output, num_items, op, init, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 410 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:90:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 90 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:122:16: note: in instantiation of function template specialization 'at::cuda::cub::unique' requested here 122 | cuda::cub::unique(data, data_out.mutable_data_ptr(), length.mutable_data_ptr(), num_inp); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:387:38: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 387 | CUB_WRAPPER(NO_ROCM(at_cuda_detail)::hipcub::DeviceSelect::Unique, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 388 | input, output, num_selected_out, num_items, at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:41:3: note: expanded from macro 'CUB_WRAPPER' 41 | func(nullptr, temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:126:20: note: in instantiation of function template specialization 'at::cuda::cub::run_length_encode' requested here 126 | at::cuda::cub::run_length_encode( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:186:12: note: in instantiation of function template specialization 'at::native::internal::(anonymous namespace)::compute_unique' requested here 186 | return compute_unique( | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:323:10: note: in instantiation of member function 'at::native::internal::UniqueCub::operator()' requested here 323 | return UniqueCub{}(*self_c, consecutive, return_inverse, return_counts); | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UniqueCub.hip:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:398:30: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 397 | CUB_WRAPPER( | ~~~~~~~~~~~~ 398 | NO_ROCM(at_cuda_detail)::hipcub::DeviceRunLengthEncode::Encode, | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 399 | input, output, counts_out, length_out, num_items, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 400 | at::hip::getCurrentHIPStreamMasqueradingAsCUDA()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/cub.cuh:44:3: note: expanded from macro 'CUB_WRAPPER' 44 | func(temp_storage.get(), temp_storage_bytes, __VA_ARGS__); \ | ^~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 64 warnings generated when compiling for host. [1563/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleBicubic2d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBicubic2d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1564/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleBilinear2d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleBilinear2d.hip:13: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1565/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_Unique.hip.o [1566/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleLinear1d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleLinear1d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1567/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleNearest1d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest1d.hip:11: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1568/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleNearest2d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest2d.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1569/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleNearest3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleNearest3d.hip:4: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSample.cuh:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1570/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_UpSampleTrilinear3d.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/UpSampleTrilinear3d.hip:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1571/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_WeightNorm.hip.o [1572/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ZetaKernel.hip.o [1573/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_ValidateCompressedIndicesKernel.hip.o [1574/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_airy_ai.hip.o [1575/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_bessel_j0.hip.o [1576/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_bessel_j1.hip.o [1577/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_bessel_y0.hip.o [1578/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_bessel_y1.hip.o [1579/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_chebyshev_polynomial_t.hip.o [1580/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_chebyshev_polynomial_u.hip.o [1581/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_chebyshev_polynomial_v.hip.o [1582/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_chebyshev_polynomial_w.hip.o [1583/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_fused_adam_amsgrad_impl.hip.o [1584/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_fused_adam_impl.hip.o [1585/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_fused_adamw_amsgrad_impl.hip.o [1586/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_fused_adamw_impl.hip.o [1587/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_hermite_polynomial_h.hip.o [1588/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_hermite_polynomial_he.hip.o [1589/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_int4mm.hip.o [1590/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_laguerre_polynomial_l.hip.o [1591/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_layer_norm_kernel.hip.o [1592/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_legendre_polynomial_p.hip.o [1593/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_modified_bessel_i0.hip.o [1594/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_modified_bessel_i1.hip.o [1595/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_modified_bessel_k0.hip.o [1596/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_modified_bessel_k1.hip.o [1597/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_scaled_modified_bessel_k0.hip.o [1598/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_group_norm_kernel.hip.o [1599/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_scaled_modified_bessel_k1.hip.o [1600/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_shifted_chebyshev_polynomial_t.hip.o [1601/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_shifted_chebyshev_polynomial_u.hip.o [1602/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_shifted_chebyshev_polynomial_v.hip.o [1603/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_shifted_chebyshev_polynomial_w.hip.o [1604/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/torch_hip_generated_spherical_bessel_j0.hip.o [1605/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/nested/hip/torch_hip_generated_NestedTensorBinaryOps.hip.o [1606/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/nested/hip/torch_hip_generated_NestedTensorMatmul.hip.o [1607/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SoftMax.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SoftMax.hip:14: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1608/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/nested/hip/torch_hip_generated_NestedTensorTransformerFunctions.hip.o [1609/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseHIPTensor.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensor.hip:10: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ 1 warning generated when compiling for host. [1610/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseHIPTensorMath.hip.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1010. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1012. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1030. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1031. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1100. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1101. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for gfx1102. In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:12: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPApplyUtils.cuh:6: In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPApplyUtils.cuh:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Atomic.cuh:338:81: warning: 'atomicAddNoRet' is deprecated: use atomicAdd instead [-Wdeprecated-declarations] 338 | static inline __device__ void gpuAtomicAddNoReturn(float *address, float val) { atomicAddNoRet(address, val); } | ^ /usr/include/hip/amd_detail/amd_hip_atomic.h:279:1: note: 'atomicAddNoRet' has been explicitly marked deprecated here 279 | DEPRECATED("use atomicAdd instead") | ^ /usr/include/hip/hip_runtime_api.h:580:41: note: expanded from macro 'DEPRECATED' 580 | #define DEPRECATED(msg) __attribute__ ((deprecated(msg))) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:318:5: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 318 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:615:7: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 615 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:720:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 720 | c10::hip::GetDevice(&curDevice); | ^~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:736:3: warning: ignoring return value of function declared with 'nodiscard' attribute [-Wunused-result] 736 | hipDeviceSynchronize(); | ^~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:894:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 894 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPTensorMath.hip:905:13: warning: 'HIPSPARSE_ORDER_COLUMN' is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 905 | HIPSPARSE_ORDER_COLUMN | ^ /usr/include/hipsparse/hipsparse.h:10736:28: note: 'HIPSPARSE_ORDER_COLUMN' has been explicitly marked deprecated here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^ /usr/include/hipsparse/hipsparse.h:66:41: note: expanded from macro 'HIPSPARSE_DEPRECATED_MSG' 66 | #define HIPSPARSE_DEPRECATED_MSG(MSG) [[deprecated(MSG)]] | ^ 13 warnings generated when compiling for host. [1611/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseMatMul.hip.o [1612/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseSemiStructuredApplyDense.hip.o [1613/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseSemiStructuredLinear.hip.o [1614/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseCsrTensorMath.hip.o [1615/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseSemiStructuredOps.hip.o [1616/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseSemiStructuredTile.hip.o [1617/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/torch_hip_generated_SparseSemiSturcturedApply.hip.o [1618/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_Activation.hip.o [1619/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_EmbeddingBag.hip.o [1620/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_AffineQuantizer.hip.o [1621/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_FusedObsFakeQuant.hip.o [1622/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_IntReprQuant.hip.o [1623/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_MakePerTensorQuantizedTensor.hip.o [1624/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/torch_hip_generated_FakeQuantizeCore.hip.o [1625/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/transformers/hip/torch_hip_generated_attention.hip.o [1626/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/transformers/hip/torch_hip_generated_attention_backward.hip.o [1627/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/torch_hip_generated_intra_node_comm.cu.o [1628/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/torch_hip_generated_Utils.cu.o [1629/1912] Linking CXX shared library lib/libshm.so Warning: Unused direct dependencies: /lib64/libprotobuf.so.30 /lib64/libm.so.6 [1630/1912] Linking CXX executable bin/torch_shm_manager Warning: Unused direct dependencies: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libshm.so /lib64/libm.so.6 [1631/1912] Building HIPCC object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/quantization/torch_hip_generated_quantization_gpu.cu.o [1632/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/nested/hip/NestedTensorTransformerUtils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorTransformerUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1633/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/nested/hip/NestedTensorTransformerFunctions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/hip/NestedTensorTransformerFunctions.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1634/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/SparseBlas.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlas.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1635/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/SparseBlasImpl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasImpl.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasImpl.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1636/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/SparseBlasLegacy.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseBlasLegacy.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1637/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp: In function ‘void at::native::sparse::cuda::{anonymous}::_csrmm2(char, char, int64_t, int64_t, int64_t, int64_t, T*, T*, int*, int*, T*, int64_t, T*, T*, int64_t, hipDataType)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:144:5: warning: ‘HIPSPARSE_ORDER_COLUMN’ is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 144 | HIPSPARSE_ORDER_COLUMN /* memory layout, ONLY column-major is supported now */ | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Exceptions.h:70:31: note: in definition of macro ‘TORCH_CUDASPARSE_CHECK’ 70 | hipsparseStatus_t __err = EXPR; \ | ^~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContextLight.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h:4: /usr/include/hipsparse/hipsparse.h:10736:5: note: declared here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:153:5: warning: ‘HIPSPARSE_ORDER_COLUMN’ is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 153 | HIPSPARSE_ORDER_COLUMN /* memory layout, ONLY column-major is supported now */ | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/Exceptions.h:70:31: note: in definition of macro ‘TORCH_CUDASPARSE_CHECK’ 70 | hipsparseStatus_t __err = EXPR; \ | ^~~~ /usr/include/hipsparse/hipsparse.h:10736:5: note: declared here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp: In instantiation of ‘void at::native::sparse::cuda::{anonymous}::_csrmm2(char, char, int64_t, int64_t, int64_t, int64_t, T*, T*, int*, int*, T*, int64_t, T*, T*, int64_t, hipDataType) [with T = float; int64_t = long int; hipDataType = hipDataType]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:217:10: required from here 217 | _csrmm2(transa, transb, m, n, k, nnz, &alpha, csrvala, csrrowptra, csrcolinda, b, ldb, &beta, c, ldc, HIP_R_32F); | ~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:158:20: warning: unused variable ‘prop’ [-Wunused-variable] 158 | hipDeviceProp_t* prop = at::cuda::getCurrentDeviceProperties(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp: In instantiation of ‘void at::native::sparse::cuda::{anonymous}::_csrmm2(char, char, int64_t, int64_t, int64_t, int64_t, T*, T*, int*, int*, T*, int64_t, T*, T*, int64_t, hipDataType) [with T = double; int64_t = long int; hipDataType = hipDataType]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:226:10: required from here 226 | _csrmm2(transa, transb, m, n, k, nnz, &alpha, csrvala, csrrowptra, csrcolinda, b, ldb, &beta, c, ldc, HIP_R_64F); | ~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:158:20: warning: unused variable ‘prop’ [-Wunused-variable] 158 | hipDeviceProp_t* prop = at::cuda::getCurrentDeviceProperties(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp: In instantiation of ‘void at::native::sparse::cuda::{anonymous}::_csrmm2(char, char, int64_t, int64_t, int64_t, int64_t, T*, T*, int*, int*, T*, int64_t, T*, T*, int64_t, hipDataType) [with T = HIP_vector_type; int64_t = long int; hipDataType = hipDataType]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:235:10: required from here 235 | _csrmm2(transa, transb, m, n, k, nnz, | ~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 236 | reinterpret_cast(&alpha), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 237 | reinterpret_cast(csrvala), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 238 | csrrowptra, | ~~~~~~~~~~~ 239 | csrcolinda, | ~~~~~~~~~~~ 240 | reinterpret_cast(b), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 241 | ldb, | ~~~~ 242 | reinterpret_cast(&beta), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 243 | reinterpret_cast(c), ldc, HIP_C_32F); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:158:20: warning: unused variable ‘prop’ [-Wunused-variable] 158 | hipDeviceProp_t* prop = at::cuda::getCurrentDeviceProperties(); | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp: In instantiation of ‘void at::native::sparse::cuda::{anonymous}::_csrmm2(char, char, int64_t, int64_t, int64_t, int64_t, T*, T*, int*, int*, T*, int64_t, T*, T*, int64_t, hipDataType) [with T = HIP_vector_type; int64_t = long int; hipDataType = hipDataType]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:252:10: required from here 252 | _csrmm2(transa, transb, m, n, k, nnz, | ~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 253 | reinterpret_cast(&alpha), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 254 | reinterpret_cast(csrvala), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 255 | csrrowptra, | ~~~~~~~~~~~ 256 | csrcolinda, | ~~~~~~~~~~~ 257 | reinterpret_cast(b), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 258 | ldb, | ~~~~ 259 | reinterpret_cast(&beta), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 260 | reinterpret_cast(c), ldc, HIP_C_64F); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/SparseHIPBlas.cpp:158:20: warning: unused variable ‘prop’ [-Wunused-variable] 158 | hipDeviceProp_t* prop = at::cuda::getCurrentDeviceProperties(); | ^~~~ [1638/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/sparse/hip/cuSPARSELtOps.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/hip/cuSPARSELtOps.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1639/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/hip/Activation.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/hip/Activation.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1640/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/BinaryOps.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1641/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/Conv.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1642/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/ConvPrepack.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1643/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/ConvUnpackImpl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1644/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/Linear.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1645/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/LinearPrepack.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1646/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/LinearUnpackImpl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1647/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/transformers/hip/sdp_utils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/hip/sdp_utils.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1648/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/CachingHostAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CachingHostAllocator.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1649/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/quantized/cudnn/hip/Pooling.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/Pooling.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cudnn/hip/Pooling.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1650/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/CuSparseHandlePool.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1651/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/EmptyTensor.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1652/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/Exceptions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1653/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/CublasHandlePool.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CublasHandlePool.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContextLight.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/CublasHandlePool.cpp:2: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1654/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/HIPContext.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1655/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/HIPGeneratorImpl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGeneratorImpl.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1656/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/HIPGraph.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGraph.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1657/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/HIPSparseBlas.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1658/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/HIPSparseDescriptors.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseDescriptors.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseDescriptors.cpp: In function ‘void* at::cuda::sparse::createRawDnMatDescriptor(const at::Tensor&, int64_t, bool)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseDescriptors.cpp:83:16: warning: ‘HIPSPARSE_ORDER_COLUMN’ is deprecated: Please use HIPSPARSE_ORDER_COL instead [-Wdeprecated-declarations] 83 | auto order = HIPSPARSE_ORDER_COLUMN; | ^~~~~~~~~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContextLight.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPSparseDescriptors.cpp:2: /usr/include/hipsparse/hipsparse.h:10736:5: note: declared here 10736 | HIPSPARSE_ORDER_COLUMN HIPSPARSE_DEPRECATED_MSG("Please use HIPSPARSE_ORDER_COL instead") | ^~~~~~~~~~~~~~~~~~~~~~ [1659/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/PeerToPeerAccess.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1660/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/HIPBlas.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp: In function ‘hipblasStatus_t rocBLASStatusToHIPStatus(rocblas_status)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:40:11: warning: enumeration value ‘rocblas_status_perf_degraded’ not handled in switch [-Wswitch] 40 | switch(error) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:40:11: warning: enumeration value ‘rocblas_status_size_query_mismatch’ not handled in switch [-Wswitch] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:40:11: warning: enumeration value ‘rocblas_status_continue’ not handled in switch [-Wswitch] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:40:11: warning: enumeration value ‘rocblas_status_check_numerics_fail’ not handled in switch [-Wswitch] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:40:11: warning: enumeration value ‘rocblas_status_excluded_from_build’ not handled in switch [-Wswitch] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:40:11: warning: enumeration value ‘rocblas_status_arch_mismatch’ not handled in switch [-Wswitch] /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp: In function ‘void at::cuda::blas::scaled_gemm(char, char, int64_t, int64_t, int64_t, const void*, const void*, int64_t, c10::ScalarType, const void*, const void*, int64_t, c10::ScalarType, const void*, c10::ScalarType, void*, const void*, int64_t, c10::ScalarType, void*, bool)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1417:16: warning: unused variable ‘fastAccuMode’ [-Wunused-variable] 1417 | const int8_t fastAccuMode = use_fast_accum ? 1 : 0; | ^~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmStridedBatchedTypeStringAndOps() [with T = double]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:285:79: required from ‘at::cuda::tunable::GemmStridedBatchedTunableOp::GemmStridedBatchedTunableOp() [with T = double; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 285 | for (auto&& [name, op] : GetRocBlasGemmStridedBatchedTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:699:96: required from ‘void at::cuda::blas::bgemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, int64_t, const DType*, int64_t, int64_t, at::opmath_type, DType*, int64_t, int64_t, int64_t) [with DType = double; int64_t = long int; at::opmath_type = double]’ 699 | static tunable::GemmStridedBatchedTunableOp bgemm{}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:723:26: required from here 723 | bgemm_tunable(CUDABLAS_BGEMM_ARGS(double)); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ In file included from /usr/include/rocblas/rocblas.h:38, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:13: /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmStridedBatchedTypeStringAndOps() [with T = float]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:285:79: required from ‘at::cuda::tunable::GemmStridedBatchedTunableOp::GemmStridedBatchedTunableOp() [with T = float; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 285 | for (auto&& [name, op] : GetRocBlasGemmStridedBatchedTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:699:96: required from ‘void at::cuda::blas::bgemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, int64_t, const DType*, int64_t, int64_t, at::opmath_type, DType*, int64_t, int64_t, int64_t) [with DType = float; int64_t = long int; at::opmath_type = float]’ 699 | static tunable::GemmStridedBatchedTunableOp bgemm{}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:734:25: required from here 734 | bgemm_tunable(CUDABLAS_BGEMM_ARGS(float)); | ~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmStridedBatchedTypeStringAndOps() [with T = c10::complex]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:285:79: required from ‘at::cuda::tunable::GemmStridedBatchedTunableOp::GemmStridedBatchedTunableOp() [with T = c10::complex; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 285 | for (auto&& [name, op] : GetRocBlasGemmStridedBatchedTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:699:96: required from ‘void at::cuda::blas::bgemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, int64_t, const DType*, int64_t, int64_t, at::opmath_type, DType*, int64_t, int64_t, int64_t) [with DType = c10::complex; int64_t = long int; at::opmath_type = c10::complex]’ 699 | static tunable::GemmStridedBatchedTunableOp bgemm{}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:745:40: required from here 745 | bgemm_tunable>(CUDABLAS_BGEMM_ARGS(c10::complex)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmStridedBatchedTypeStringAndOps() [with T = c10::complex]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:285:79: required from ‘at::cuda::tunable::GemmStridedBatchedTunableOp::GemmStridedBatchedTunableOp() [with T = c10::complex; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 285 | for (auto&& [name, op] : GetRocBlasGemmStridedBatchedTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:699:96: required from ‘void at::cuda::blas::bgemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, int64_t, const DType*, int64_t, int64_t, at::opmath_type, DType*, int64_t, int64_t, int64_t) [with DType = c10::complex; int64_t = long int; at::opmath_type = c10::complex]’ 699 | static tunable::GemmStridedBatchedTunableOp bgemm{}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:756:39: required from here 756 | bgemm_tunable>(CUDABLAS_BGEMM_ARGS(c10::complex)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmStridedBatchedTypeStringAndOps() [with T = c10::Half]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:285:79: required from ‘at::cuda::tunable::GemmStridedBatchedTunableOp::GemmStridedBatchedTunableOp() [with T = c10::Half; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 285 | for (auto&& [name, op] : GetRocBlasGemmStridedBatchedTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:699:96: required from ‘void at::cuda::blas::bgemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, int64_t, const DType*, int64_t, int64_t, at::opmath_type, DType*, int64_t, int64_t, int64_t) [with DType = c10::Half; int64_t = long int; at::opmath_type = float]’ 699 | static tunable::GemmStridedBatchedTunableOp bgemm{}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:767:28: required from here 767 | bgemm_tunable(CUDABLAS_BGEMM_ARGS(at::Half)); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmStridedBatchedTypeStringAndOps() [with T = c10::BFloat16]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:285:79: required from ‘at::cuda::tunable::GemmStridedBatchedTunableOp::GemmStridedBatchedTunableOp() [with T = c10::BFloat16; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 285 | for (auto&& [name, op] : GetRocBlasGemmStridedBatchedTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:699:96: required from ‘void at::cuda::blas::bgemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, int64_t, const DType*, int64_t, int64_t, at::opmath_type, DType*, int64_t, int64_t, int64_t) [with DType = c10::BFloat16; int64_t = long int; at::opmath_type = float]’ 699 | static tunable::GemmStridedBatchedTunableOp bgemm{}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:778:32: required from here 778 | bgemm_tunable(CUDABLAS_BGEMM_ARGS(at::BFloat16)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:249:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 249 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 250 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 251 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 252 | compute_type, | ~~~~~~~~~~~~~ 253 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 254 | nullptr, | ~~~~~~~~ 255 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:258:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 258 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 259 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 260 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 261 | compute_type, | ~~~~~~~~~~~~~ 262 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 263 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 264 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmTypeStringAndOps() [with T = double]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:242:65: required from ‘at::cuda::tunable::GemmTunableOp::GemmTunableOp() [with T = double; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 242 | for (auto&& [name, op] : GetRocBlasGemmTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1095:82: required from ‘void at::cuda::blas::gemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, const DType*, int64_t, at::opmath_type, DType*, int64_t) [with DType = double; int64_t = long int; at::opmath_type = double]’ 1095 | static tunable::GemmTunableOp gemm{}; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1119:25: required from here 1119 | gemm_tunable(CUDABLAS_GEMM_ARGS(double)); | ~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmTypeStringAndOps() [with T = float]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:242:65: required from ‘at::cuda::tunable::GemmTunableOp::GemmTunableOp() [with T = float; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 242 | for (auto&& [name, op] : GetRocBlasGemmTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1095:82: required from ‘void at::cuda::blas::gemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, const DType*, int64_t, at::opmath_type, DType*, int64_t) [with DType = float; int64_t = long int; at::opmath_type = float]’ 1095 | static tunable::GemmTunableOp gemm{}; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1130:24: required from here 1130 | gemm_tunable(CUDABLAS_GEMM_ARGS(float)); | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmTypeStringAndOps() [with T = c10::complex]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:242:65: required from ‘at::cuda::tunable::GemmTunableOp::GemmTunableOp() [with T = c10::complex; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 242 | for (auto&& [name, op] : GetRocBlasGemmTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1095:82: required from ‘void at::cuda::blas::gemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, const DType*, int64_t, at::opmath_type, DType*, int64_t) [with DType = c10::complex; int64_t = long int; at::opmath_type = c10::complex]’ 1095 | static tunable::GemmTunableOp gemm{}; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1141:39: required from here 1141 | gemm_tunable>(CUDABLAS_GEMM_ARGS(c10::complex)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmTypeStringAndOps() [with T = c10::complex]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:242:65: required from ‘at::cuda::tunable::GemmTunableOp::GemmTunableOp() [with T = c10::complex; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 242 | for (auto&& [name, op] : GetRocBlasGemmTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1095:82: required from ‘void at::cuda::blas::gemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, const DType*, int64_t, at::opmath_type, DType*, int64_t) [with DType = c10::complex; int64_t = long int; at::opmath_type = c10::complex]’ 1095 | static tunable::GemmTunableOp gemm{}; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1152:38: required from here 1152 | gemm_tunable>(CUDABLAS_GEMM_ARGS(c10::complex)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmTypeStringAndOps() [with T = c10::Half]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:242:65: required from ‘at::cuda::tunable::GemmTunableOp::GemmTunableOp() [with T = c10::Half; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 242 | for (auto&& [name, op] : GetRocBlasGemmTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1095:82: required from ‘void at::cuda::blas::gemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, const DType*, int64_t, at::opmath_type, DType*, int64_t) [with DType = c10::Half; int64_t = long int; at::opmath_type = float]’ 1095 | static tunable::GemmTunableOp gemm{}; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1163:27: required from here 1163 | gemm_tunable(CUDABLAS_GEMM_ARGS(at::Half)); | ~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h: In instantiation of ‘auto at::cuda::tunable::GetRocBlasGemmTypeStringAndOps() [with T = c10::BFloat16]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/TunableGemm.h:242:65: required from ‘at::cuda::tunable::GemmTunableOp::GemmTunableOp() [with T = c10::BFloat16; at::cuda::tunable::BlasOp ALayout = at::cuda::tunable::BlasOp::T; at::cuda::tunable::BlasOp BLayout = at::cuda::tunable::BlasOp::T]’ 242 | for (auto&& [name, op] : GetRocBlasGemmTypeStringAndOps()) { | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1095:82: required from ‘void at::cuda::blas::gemm_tunable(char, char, int64_t, int64_t, int64_t, at::opmath_type, const DType*, int64_t, const DType*, int64_t, at::opmath_type, DType*, int64_t) [with DType = c10::BFloat16; int64_t = long int; at::opmath_type = float]’ 1095 | static tunable::GemmTunableOp gemm{}; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPBlas.cpp:1174:31: required from here 1174 | gemm_tunable(CUDABLAS_GEMM_ARGS(at::BFloat16)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:179:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 179 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 180 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 181 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 182 | compute_type, | ~~~~~~~~~~~~~ 183 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 184 | nullptr, | ~~~~~~~~ 185 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:188:60: warning: ‘rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle, rocblas_datatype, rocblas_datatype, rocblas_datatype, uint32_t, rocblas_int*, rocblas_int*)’ is deprecated: "rocblas_gemm_ex_get_solutions_by_type is a beta feature and is subject to " "change in future releases" [-Wdeprecated-declarations] 188 | TORCH_ROCBLAS_CHECK(rocblas_gemm_ex_get_solutions_by_type(handle, | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ 189 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 190 | input_output_type, | ~~~~~~~~~~~~~~~~~~ 191 | compute_type, | ~~~~~~~~~~~~~ 192 | rocblas_gemm_flags_none, | ~~~~~~~~~~~~~~~~~~~~~~~~ 193 | solutions.data(), | ~~~~~~~~~~~~~~~~~ 194 | &solution_size)); | ~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/tunable/GemmRocblas.h:17:28: note: in definition of macro ‘TORCH_ROCBLAS_CHECK’ 17 | rocblas_status __err = EXPR; \ | ^~~~ /usr/include/rocblas/internal/rocblas-beta.h:217:31: note: declared here 217 | ROCBLAS_EXPORT rocblas_status rocblas_gemm_ex_get_solutions_by_type(rocblas_handle handle, | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1661/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/detail/HIPHooks.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/HIPHooks.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/HIPHooks.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h: At global scope: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /usr/include/hip/hip_runtime.h:70, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPGeneratorImpl.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/detail/HIPHooks.cpp:4: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1662/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/impl/HIPCachingAllocatorMasqueradingAsCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1663/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/impl/HIPGuardImplMasqueradingAsCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1664/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/PinnedMemoryAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/PinnedMemoryAllocator.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1665/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/llvm_basic.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1666/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/llvm_complex.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1667/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/tunable/StreamTimer.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1668/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/hip/tunable/Tunable.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1669/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Activation.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Activation.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1670/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Distributions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Distributions.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1671/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Equal.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Equal.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1672/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/GridSampler.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/GridSampler.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1673/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/IndexKernel.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/IndexKernel.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1674/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/LinearAlgebraStubs.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/LinearAlgebraStubs.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1675/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/ReduceOps.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOpsUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ReduceOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1676/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Resize.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Resize.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Resize.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/EmptyTensor.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Resize.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1677/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/ScanKernels.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/ScanKernels.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1678/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Sort.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sort.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1679/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Sorting.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Sorting.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1680/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/Blas.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Blas.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/Blas.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1681/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/TensorCompare.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1682/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/SpectralOps.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SpectralOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SpectralOps.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h: At global scope: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContextLight.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/SpectralOps.cpp:4: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1683/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/TensorModeKernel.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOpsUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorModeKernel.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1684/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/TensorTopK.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorTopK.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1685/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/TensorShapeHIP.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/TensorShapeHIP.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1686/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/jit_utils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/jit_utils.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /usr/include/hip/hip_runtime.h:70, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/jit_utils.cpp:2: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1687/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/linalg/BatchLinearAlgebra.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebra.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1688/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLib.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLib.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1689/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/linalg/CusolverDnHandlePool.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1690/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLibBlas.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/hip/linalg/BatchLinearAlgebraLibBlas.cpp:25: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1691/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/hip/linalg/HIPSolver.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1692/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/RegisterNestedTensorCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorCUDA.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorCUDA.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1693/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/RegisterQuantizedCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedCUDA.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedCUDA.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1694/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/RegisterSparseCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCUDA.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCUDA.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1695/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/RegisterSparseCsrCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrCUDA.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrCUDA.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1696/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/miopen/BatchNorm_miopen.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/BatchNorm_miopen.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1697/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/miopen/Conv_miopen.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/Conv_miopen.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1698/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/miopen/RNN_miopen.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/Exception.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Device.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RNN.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp: In function ‘void at::native::_viewOrCopyParams(at::MatrixRef, at::MatrixRef, bool)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:271:40: warning: ‘at::DeprecatedTypeProperties& at::Tensor::type() const’ is deprecated: Tensor.type() is deprecated. Instead use Tensor.options(), which in many cases (e.g. in a constructor) is a drop-in replacement. If you were using data from type(), that is now available from Tensor itself, so instead of tensor.type().scalar_type(), use tensor.scalar_type() instead and instead of tensor.type().backend() use tensor.device(). [-Wdeprecated-declarations] 271 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/macros/Macros.h:204:64: note: in definition of macro ‘C10_UNLIKELY’ 204 | #define C10_UNLIKELY(expr) (__builtin_expect(static_cast(expr), 0)) | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/Exception.h:534:7: note: in expansion of macro ‘C10_UNLIKELY_OR_CONST’ 534 | if (C10_UNLIKELY_OR_CONST(!(cond))) { \ | ^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:271:13: note: in expansion of macro ‘TORCH_CHECK’ 271 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:225:30: note: declared here 225 | DeprecatedTypeProperties & type() const { | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:271:59: warning: ‘at::DeprecatedTypeProperties& at::Tensor::type() const’ is deprecated: Tensor.type() is deprecated. Instead use Tensor.options(), which in many cases (e.g. in a constructor) is a drop-in replacement. If you were using data from type(), that is now available from Tensor itself, so instead of tensor.type().scalar_type(), use tensor.scalar_type() instead and instead of tensor.type().backend() use tensor.device(). [-Wdeprecated-declarations] 271 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/macros/Macros.h:204:64: note: in definition of macro ‘C10_UNLIKELY’ 204 | #define C10_UNLIKELY(expr) (__builtin_expect(static_cast(expr), 0)) | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/Exception.h:534:7: note: in expansion of macro ‘C10_UNLIKELY_OR_CONST’ 534 | if (C10_UNLIKELY_OR_CONST(!(cond))) { \ | ^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:271:13: note: in expansion of macro ‘TORCH_CHECK’ 271 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:225:30: note: declared here 225 | DeprecatedTypeProperties & type() const { | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp: In function ‘void at::native::_copyParams_and_permute(at::MatrixRef, at::MatrixRef, int64_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:290:40: warning: ‘at::DeprecatedTypeProperties& at::Tensor::type() const’ is deprecated: Tensor.type() is deprecated. Instead use Tensor.options(), which in many cases (e.g. in a constructor) is a drop-in replacement. If you were using data from type(), that is now available from Tensor itself, so instead of tensor.type().scalar_type(), use tensor.scalar_type() instead and instead of tensor.type().backend() use tensor.device(). [-Wdeprecated-declarations] 290 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/macros/Macros.h:204:64: note: in definition of macro ‘C10_UNLIKELY’ 204 | #define C10_UNLIKELY(expr) (__builtin_expect(static_cast(expr), 0)) | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/Exception.h:534:7: note: in expansion of macro ‘C10_UNLIKELY_OR_CONST’ 534 | if (C10_UNLIKELY_OR_CONST(!(cond))) { \ | ^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:290:13: note: in expansion of macro ‘TORCH_CHECK’ 290 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:225:30: note: declared here 225 | DeprecatedTypeProperties & type() const { | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:290:59: warning: ‘at::DeprecatedTypeProperties& at::Tensor::type() const’ is deprecated: Tensor.type() is deprecated. Instead use Tensor.options(), which in many cases (e.g. in a constructor) is a drop-in replacement. If you were using data from type(), that is now available from Tensor itself, so instead of tensor.type().scalar_type(), use tensor.scalar_type() instead and instead of tensor.type().backend() use tensor.device(). [-Wdeprecated-declarations] 290 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/macros/Macros.h:204:64: note: in definition of macro ‘C10_UNLIKELY’ 204 | #define C10_UNLIKELY(expr) (__builtin_expect(static_cast(expr), 0)) | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/Exception.h:534:7: note: in expansion of macro ‘C10_UNLIKELY_OR_CONST’ 534 | if (C10_UNLIKELY_OR_CONST(!(cond))) { \ | ^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/miopen/RNN_miopen.cpp:290:13: note: in expansion of macro ‘TORCH_CHECK’ 290 | TORCH_CHECK(param_from.type() == param_to.type(), "parameter types mismatch"); | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:225:30: note: declared here 225 | DeprecatedTypeProperties & type() const { | ^~~~ [1699/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/AffineGridGenerator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/AffineGridGenerator.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1700/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/BatchNorm.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1701/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/ConvPlaceholders.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1702/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/ConvShared.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1703/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/Conv_v7.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1704/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/Conv_v8.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1705/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/GridSampler.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/GridSampler.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1706/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/LossCTC.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_cudnn_ctc_loss.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/LossCTC.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1707/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/MHA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/MHA.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1708/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/native/cudnn/hip/RNN.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_cudnn_init_dropout_state.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cudnn/hip/RNN.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1709/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/miopen/AutocastRNN.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/miopen/AutocastRNN.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/miopen/AutocastRNN.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1710/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/miopen/Descriptors.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/miopen/Descriptors.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1711/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/miopen/Handle.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1712/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/RegisterCUDA.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCUDA.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCUDA.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1713/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/CudaIPCTypes.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1714/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/aten/src/ATen/miopen/Types.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/miopen/Types.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1715/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/cuda/memory_snapshot.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/memory_snapshot.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1716/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/cuda/comm.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/comm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/comm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1717/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/inductor/aoti_runner/model_container_runner_cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1718/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/inductor/aoti_torch/shim_cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/shim_cuda.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1719/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/profiler/stubs/cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/util.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/stubs/cuda.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1720/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/interface.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/compiler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h: At global scope: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /usr/include/hip/hip_runtime.h:70, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/codegen/fuser/cuda/fused_kernel.h:7: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1721/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/jit/passes/frozen_conv_add_relu_fusion_cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/constants.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion_cuda.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/frozen_conv_add_relu_fusion_cuda.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1722/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/autograd/functions/comm.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/comm.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/comm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1723/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/jit/runtime/register_cuda_ops.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_cuda_ops.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/utils.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1724/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/jit/tensorexpr/cuda_codegen.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h: At global scope: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /usr/include/hip/hip_runtime.h:70, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.cpp:1: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [1725/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/NCCLUtils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1726/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/ProcessGroupCudaP2P.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1727/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/ProcessGroupNCCL.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1728/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/ProcessGroupUCC.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1729/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/UCCTracing.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1730/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/UCCUtils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1731/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/reducer_cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_kineto.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer_timer.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/reducer_cuda.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1732/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/c10d/intra_node_comm.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/intra_node_comm.hpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1733/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/distributed/rpc/tensorpipe_cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PrefixStore.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_agent.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_cuda.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/tensorpipe_cuda.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1734/1912] Building CXX object caffe2/CMakeFiles/torch.dir/__/empty.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1735/1912] Building CXX object caffe2/CMakeFiles/torch_hip.dir/__/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1736/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_functions_0.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_0.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1737/1912] Linking CXX shared library lib/libtorch_hip.so Warning: Unused direct dependencies: /lib64/libhiprtc.so.6 /lib64/libhiprand.so.1 [1738/1912] Linking CXX shared library lib/libtorch.so Warning: Unused direct dependencies: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch_cpu.so /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch_hip.so [1739/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_functions_1.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_1.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1740/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_functions_2.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_2.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1741/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_functions_3.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_3.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1742/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_functions_4.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_4.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1743/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_variable_methods.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_variable_methods.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_variable_methods.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1744/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_torch_functions_0.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_0.cpp:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1745/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_torch_functions_1.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_1.cpp:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1746/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_torch_functions_2.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_2.cpp:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1747/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_nn_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_nn_functions.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1748/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_fft_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_fft_functions.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1749/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_linalg_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_linalg_functions.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1750/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_nested_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_nested_functions.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1751/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_sparse_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_sparse_functions.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1752/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_return_types.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1753/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_special_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_special_functions.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1754/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/generated/python_enum_tag.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1755/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/DataLoader.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_numbers.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DataLoader.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1756/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Dtype.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_numbers.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Dtype.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1757/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Device.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Device.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1758/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/DynamicTypes.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/DynamicTypes.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1759/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Exceptions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Exceptions.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1760/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Layout.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1761/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/MemoryFormat.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1762/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Generator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Generator.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Generator.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1763/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/QScheme.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1764/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/PyInterpreter.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonFallbackKernel.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/PyInterpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1765/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/python_dimname.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1766/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Module.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Module.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Module.cpp:72: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1767/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Size.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1768/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Storage.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Storage.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1769/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/StorageMethods.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageMethods.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1770/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/StorageSharing.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/StorageSharing.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1771/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Stream.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Stream.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1772/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/Event.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Event.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1773/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/TypeInfo.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/TypeInfo.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/TypeInfo.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1774/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/functions/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CachedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/accumulate_grad.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/functions/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1775/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/api/src/python/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Variadic.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/variadic.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/python.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/src/python/init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/detail/static.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1776/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/profiler_python.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/profiler_python.cpp:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1777/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_anomaly_mode.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_anomaly_mode.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1778/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/init.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/init.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1779/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_saved_variable_hooks.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_saved_variable_hooks.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_saved_variable_hooks.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1780/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_cpp_function.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1781/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_engine.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_engine.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_engine.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1782/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_function.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/custom_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_function.cpp:30: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1783/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_hook.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_hook.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1784/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_legacy_variable.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_legacy_variable.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1785/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_nested_functions_manual.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nested_functions_manual.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/all.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/torch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_nested_functions_manual.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1786/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_torch_functions_manual.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_torch_functions_manual.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_torch_functions_manual.cpp:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1787/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_variable_indexing.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable_indexing.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1788/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/autograd/python_variable.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.cpp:25: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1789/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/cache_entry.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1790/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/cpp_shim.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/cpp_shim.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1791/1912] Building C object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/cpython_defs.c.o cc1: warning: command-line option ‘-std=c++17’ is valid for C++/ObjC++ but not for C [1792/1912] Building C object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/eval_frame.c.o cc1: warning: command-line option ‘-std=c++17’ is valid for C++/ObjC++ but not for C [1793/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/extra_state.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1794/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/python_compiled_autograd.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/engine.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/python_compiled_autograd.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/python_compiled_autograd.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1795/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1796/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/functorch/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/functorch/init.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1797/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/dynamo/guards.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/wrap_outputs.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/all.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/extension.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp: In member function ‘virtual bool torch::dynamo::{anonymous}::GlobalWeakRefGuardAccessor::check_nopybind(PyObject*)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp:3142:38: warning: ‘PyObject* PyWeakref_GetObject(PyObject*)’ is deprecated [-Wdeprecated-declarations] 3142 | PyObject* x = PyWeakref_GetObject(weakref); // borrowed ref | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~ In file included from /usr/include/python3.13/Python.h:113, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/python_headers.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/utils/wrap_outputs.h:8: /usr/include/python3.13/weakrefobject.h:30:44: note: declared here 30 | Py_DEPRECATED(3.13) PyAPI_FUNC(PyObject *) PyWeakref_GetObject(PyObject *ref); | ^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp: In member function ‘virtual torch::dynamo::{anonymous}::GuardDebugInfo torch::dynamo::{anonymous}::GlobalWeakRefGuardAccessor::check_verbose_nopybind(PyObject*)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/dynamo/guards.cpp:3163:38: warning: ‘PyObject* PyWeakref_GetObject(PyObject*)’ is deprecated [-Wdeprecated-declarations] 3163 | PyObject* x = PyWeakref_GetObject(weakref); // borrowed ref | ~~~~~~~~~~~~~~~~~~~^~~~~~~~~ /usr/include/python3.13/weakrefobject.h:30:44: note: declared here 30 | Py_DEPRECATED(3.13) PyAPI_FUNC(PyObject *) PyWeakref_GetObject(PyObject *ref); | ^~~~~~~~~~~~~~~~~~~ [1798/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/mtia/Module.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mtia/Module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1799/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/mps/Module.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/mps/Module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1800/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/inductor/aoti_runner/pybind.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/utils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_runner/pybind.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1801/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/inductor/aoti_eager/kernel_meta_info.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_meta_info.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_meta_info.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1802/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/inductor/aoti_eager/kernel_holder.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_eager/kernel_holder.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1803/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/backends/backend_init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1804/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1805/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/cast_all_constant_to_floating.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/cast_all_constant_to_floating.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/cast_all_constant_to_floating.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1806/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_init.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/init.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1807/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/deduplicate_initializers.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/deduplicate_initializers.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/deduplicate_initializers.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1808/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/constant_fold.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_fold.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_fold.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1809/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/eval_peephole.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eval_peephole.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eval_peephole.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1810/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/eliminate_unused_items.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eliminate_unused_items.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/eliminate_unused_items.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1811/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/constant_map.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_map.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/constant_map.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1812/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/list_model_parameters.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/list_model_parameters.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1813/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/function_substitution.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_substitution.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_substitution.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1814/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/helper.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1815/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/fixup_onnx_controlflow.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/fixup_onnx_controlflow.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1816/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/preprocess_for_onnx.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/preprocess_for_onnx.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/preprocess_for_onnx.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1817/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/peephole.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/peephole.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/peephole.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1818/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/prepare_division_for_onnx.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/prepare_division_for_onnx.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/prepare_division_for_onnx.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1819/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/scalar_type_analysis.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/scalar_type_analysis.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1820/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/remove_inplace_ops_for_onnx.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/helper.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/remove_inplace_ops_for_onnx.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1821/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/unpack_quantized_weights.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/unpack_quantized_weights.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/unpack_quantized_weights.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1822/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/function_extraction.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_extraction.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/function_extraction.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1823/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/onnx_log.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1824/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/naming.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/naming.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/naming.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1825/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/shape_type_inference.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1826/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/autograd_function_process.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1827/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/pattern_conversion/common.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/common.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/common.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1828/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/pybind_utils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/graph_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp: In function ‘pybind11::object torch::jit::toPyObject(c10::IValue)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:631:21: warning: redundant move in return statement [-Wredundant-move] 631 | return std::move(t); | ~~~~~~~~~^~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:631:21: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:666:23: warning: redundant move in return statement [-Wredundant-move] 666 | return std::move(t); | ~~~~~~~~~^~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:666:23: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:679:21: warning: redundant move in return statement [-Wredundant-move] 679 | return std::move(py_dict); | ~~~~~~~~~^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.cpp:679:21: note: remove ‘std::move’ call [1829/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/pattern_conversion/pattern_encapsulation.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_encapsulation.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1830/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/passes/onnx/pattern_conversion/pattern_conversion.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/pattern_conversion/pattern_conversion.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1831/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_arg_flatten.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.cpp: In function ‘pybind11::object torch::jit::python::{anonymous}::cast_dict(std::vector)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.cpp:130:19: warning: redundant move in return statement [-Wredundant-move] 130 | return std::move(sequence); | ~~~~~~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.cpp:130:19: note: remove ‘std::move’ call [1832/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_custom_class.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_custom_class.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1833/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_dict.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_dict.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_dict.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1834/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_interpreter.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_interpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_interpreter.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_interpreter.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1835/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_list.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_list.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_list.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1836/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_ir.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_ir.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1837/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_tracer.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tracer.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/dead_code_elimination.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tracer.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1838/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/frontend/concrete_module_type.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1839/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/frontend/tree_views.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1840/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/script_init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/ir_emitter.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/script_init.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1841/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_sugared_value.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/concrete_module_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_sugared_value.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1842/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/python/python_tree_views.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_tree_views.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1843/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/runtime/static/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/init.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/static/init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1844/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/monitor/python_init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/monitor/python_init.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1845/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/multiprocessing/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1846/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/tensorexpr/tensorexpr_init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensorexpr_init.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/tensorexpr_init.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h: At global scope: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:113:5: warning: ‘hipError_t hipCtxGetCurrent(ihipCtx_t**)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 113 | _(hipCtxGetCurrent) \ | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContextLight.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/HIPContext.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/tensorexpr/cuda_codegen.h:7: /usr/include/hip/hip_runtime_api.h:5156:12: note: declared here 5156 | hipError_t hipCtxGetCurrent(hipCtx_t* ctx); | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:116:5: warning: ‘hipError_t hipDevicePrimaryCtxGetState(hipDevice_t, unsigned int*, int*)’ is deprecated: This API is marked as deprecated and may not be supported in future releases. For more details please refer https://github.com/ROCm/HIP/blob/develop/docs/reference/deprecated_api_list.md [-Wdeprecated-declarations] 116 | _(hipDevicePrimaryCtxGetState) | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:121:39: note: in definition of macro ‘CREATE_MEMBER’ 121 | #define CREATE_MEMBER(name) decltype(&name) name; | ^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/hip/nvrtc_stub/ATenNVRTC.h:122:3: note: in expansion of macro ‘AT_FORALL_NVRTC’ 122 | AT_FORALL_NVRTC(CREATE_MEMBER) | ^~~~~~~~~~~~~~~ /usr/include/hip/hip_runtime_api.h:5359:12: note: declared here 5359 | hipError_t hipDevicePrimaryCtxGetState(hipDevice_t dev, unsigned int* flags, int* active); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~ [1847/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/onnx/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/init.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/python_arg_flatten.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/onnx/shape_type_inference.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/onnx/init.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1848/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/profiler/python/combined_traceback.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/interpreter.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/combined_traceback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/combined_traceback.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/combined_traceback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1849/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/profiler/python/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/orchestration/observer.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/api.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/kineto_shim.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/collection.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/profiler/python/init.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1850/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/serialization.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/serialization.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1851/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/tensor/python_tensor.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/tensor/python_tensor.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1852/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1853/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/throughput_benchmark.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/throughput_benchmark.h:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1854/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/device_lazy_init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1855/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1856/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/invalid_arguments.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1857/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/object_ptr.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1858/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/nested.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/all.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/torch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/nested.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1859/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/python_arg_parser.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1860/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/python_symnode.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_symnode.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_symnode.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1861/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/pybind.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/pybind.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1862/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/pyobject_preservation.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1863/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/structseq.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1864/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/python_dispatch.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_dispatch.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1865/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_dtypes.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1866/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_apply.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_apply.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1867/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_layouts.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1868/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_memoryformats.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1869/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_qschemes.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1870/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_list.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_list.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1871/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_new.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/python_arg_parser.h:55, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_new.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_new.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1872/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_types.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_types.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1873/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/tensor_numpy.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/tensor_numpy.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1874/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/verbose.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1875/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/utils/disable_torch_function.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_variable.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/utils/disable_torch_function.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1876/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cpu/Module.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1877/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/lazy/python/python_util.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/named_value.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/passes/symbolic_shape_analysis.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/shape.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/backend/backend_data.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/tensor.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/core/debug_util.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/python_util.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1878/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/lazy/python/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/python/init.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1879/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/Event.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Event.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1880/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/Module.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Module.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1881/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/python_comm.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/THCP.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/python_comm.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1882/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/Stream.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Stream.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1883/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/Graph.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/Graph.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/operator.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/ir/ir.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/function_impl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/method.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/object.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/api/module.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1884/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/shared/cudart.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1885/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/shared/nvtx.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1886/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/CUDAPluggableAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1887/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/shared/cudnn.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1888/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/cuda/utils.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/Size.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/THP.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/THCP.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/cuda/utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [1889/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/c10d/python_comm_hook.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Types.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Backend.hpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/ProcessGroup.hpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/comm.hpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/python_comm_hook.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1890/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/autograd/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/graph_task.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/function.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/python_cpp_function.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/init.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/init.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1891/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/c10d/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/Store.hpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/FileStore.hpp:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/init.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/PyProcessGroup.hpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/c10d/init.cpp:19: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1892/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/profiler/remote_profiler_manager.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/init.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/init.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/init.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1893/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/py_rref.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_impl.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/py_rref.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1894/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/python_functions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_functions.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1895/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/python_rpc_handler.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1896/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/request_callback_impl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.cpp:22: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1897/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/testing/init.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/testing/init.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/request_callback_impl.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1898/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/unpickled_python_call.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_call.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1899/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/distributed/rpc/unpickled_python_remote_call.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_command_base.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/python_rpc_handler.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/unpickled_python_remote_call.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call [1900/1912] Building CXX object caffe2/torch/CMakeFiles/torch_python.dir/csrc/jit/runtime/register_distributed_ops.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_distributed_ops.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/types.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/message.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rpc_agent.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/functions/recvrpc_backward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/context/container.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_distributed_ops.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/script_type_parser.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/unpickler.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/pickle.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/utils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/rref_context.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/distributed/rpc/torchscript_functions.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_distributed_ops.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h: In function ‘torch::jit::Expr torch::jit::pep604union_to_union(const Expr&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: warning: redundant move in return statement [-Wredundant-move] 1263 | return std::move(synthesised_union); | ~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/tree_views.h:1263:19: note: remove ‘std::move’ call In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/runtime/register_distributed_ops.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1901/1912] Building CXX object caffe2/torch/CMakeFiles/nnapi_backend.dir/csrc/jit/backends/nnapi/nnapi_backend_lib.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_lib.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1902/1912] Linking CXX shared library lib/libtorch_python.so Warning: Unused direct dependencies: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch.so [1903/1912] Building CXX object functorch/CMakeFiles/functorch.dir/csrc/dim/dim.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [1904/1912] Building C object functorch/CMakeFiles/functorch.dir/csrc/dim/dim_opcode.c.o cc1: warning: command-line option ‘-std=c++17’ is valid for C++/ObjC++ but not for C [1905/1912] Building CXX object caffe2/torch/CMakeFiles/nnapi_backend.dir/csrc/jit/backends/nnapi/nnapi_backend_preprocess.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_preprocess.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend_interface.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/backend.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/backends/nnapi/nnapi_backend_preprocess.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h: In function ‘pybind11::object torch::jit::createPyObjectForStack(Stack&&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: warning: redundant move in return statement [-Wredundant-move] 1078 | return std::move(return_values); | ~~~~~~~~~^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/python/pybind_utils.h:1078:19: note: remove ‘std::move’ call [1906/1912] Linking CXX shared library lib/libnnapi_backend.so Warning: Unused direct dependencies: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch.so /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch_hip.so [1907/1912] Building CXX object functorch/CMakeFiles/functorch.dir/csrc/init_dim_only.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/variable.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/all.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/extension.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/functorch/csrc/init_dim_only.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/autograd_not_implemented_fallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/api/include/torch/autograd.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [1908/1912] Linking CXX shared module functorch/functorch.so Warning: Unused direct dependencies: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch_python.so /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch.so /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch_cpu.so /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib/libtorch_hip.so [1908/1912] Install the project... -- Install configuration: "RelWithDebInfo" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libc10_hip.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libc10.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libcaffe2_nvrtc.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libshm.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/bin/torch_shm_manager" to "$ORIGIN/../lib" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libtorch_python.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libtorch_cpu.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libtorch_hip.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libtorch.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib/libtorch_global_deps.so" to "$ORIGIN" -- Set non-toolchain portion of runtime path of "/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/functorch/functorch.so" to "$ORIGIN/../torch/lib" /usr/lib/python3.13/site-packages/setuptools/dist.py:318: InformationOnly: Normalizing '2.4.0a0+gitUnknown' to '2.4.0a0+gitunknown' self.metadata.version = self._normalize_version(self.metadata.version) Building wheel torch-2.4.0a0+gitUnknown -- Building version 2.4.0a0+gitUnknown cmake3 -GNinja -DBUILD_CUSTOM_PROTOBUF=OFF -DBUILD_NVFUSER=OFF -DBUILD_PYTHON=True -DBUILD_SHARED_LIBS=ON -DBUILD_TEST=False -DCMAKE_BUILD_TYPE=RelWithDebInfo -DCMAKE_EXE_LINKER_FLAGS=-pie -DCMAKE_FIND_PACKAGE_PREFER_CONFIG=ON -DCMAKE_INSTALL_PREFIX=/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch -DCMAKE_PREFIX_PATH=/usr/lib/python3.13/site-packages -DPython_EXECUTABLE=/usr/bin/python3 -DTORCH_BUILD_VERSION=2.4.0a0+gitUnknown -DUSE_CUDA=OFF -DUSE_DISTRIBUTED=ON -DUSE_FAKELOWP=OFF -DUSE_FBGEMM=OFF -DUSE_FLASH_ATTENTION=OFF -DUSE_GLOO=OFF -DUSE_GOLD_LINKER=ON -DUSE_ITT=OFF -DUSE_KINETO=OFF -DUSE_LITE_INTERPRETER_PROFILER=OFF -DUSE_LITE_PROTO=OFF -DUSE_MAGMA=ON -DUSE_MEM_EFF_ATTENTION=OFF -DUSE_MKLDNN=OFF -DUSE_MPI=ON -DUSE_NCCL=OFF -DUSE_NNPACK=OFF -DUSE_NUMPY=True -DUSE_OPENMP=ON -DUSE_PYTORCH_QNNPACK=OFF -DUSE_ROCM=ON -DUSE_SYSTEM_CPUINFO=ON -DUSE_SYSTEM_EIGEN_INSTALL=ON -DUSE_SYSTEM_FP16=ON -DUSE_SYSTEM_FXDIV=ON -DUSE_SYSTEM_LIBS=OFF -DUSE_SYSTEM_ONNX=ON -DUSE_SYSTEM_PSIMD=ON -DUSE_SYSTEM_PTHREADPOOL=ON -DUSE_SYSTEM_PYBIND11=OFF -DUSE_SYSTEM_SLEEF=ON -DUSE_SYSTEM_XNNPACK=ON -DUSE_TENSORPIPE=ON -DUSE_XNNPACK=ON -DUSE_XPU=OFF /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0 cmake3 --build . --target install --config RelWithDebInfo -- -j 2 running build running build_py creating build/lib.linux-x86_64-cpython-313 creating build/lib.linux-x86_64-cpython-313/functorch copying functorch/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch creating build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/yaml_utils.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/native_function_generation.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen_functionalization_type.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/utils.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen_aoti_c_shim.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/local.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/model.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/context.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen_executorch.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/code_template.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen_lazy_tensor.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen_backend_stubs.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/gen_vmap_plumbing.py -> build/lib.linux-x86_64-cpython-313/torchgen copying torchgen/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen creating build/lib.linux-x86_64-cpython-313/torch copying torch/overrides.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_tensor_str.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/random.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/return_types.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_namedtensor_internals.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/library.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_lowrank.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/storage.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_weights_only_unpickler.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_streambase.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_sources.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_storage_docs.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_vmap_internals.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_jit_internal.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/torch_version.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_size_docs.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_linalg_utils.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/__future__.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_ops.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_deploy.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_utils_internal.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/types.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_tensor_docs.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_tensor.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/serialization.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_guards.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/quasirandom.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/__config__.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_python_dispatcher.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_utils.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_VF.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/hub.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_appdirs.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_custom_ops.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/functional.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_lobpcg.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/__init__.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_meta_registrations.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_classes.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_torch_docs.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/_compile.py -> build/lib.linux-x86_64-cpython-313/torch copying torch/version.py -> build/lib.linux-x86_64-cpython-313/torch creating build/lib.linux-x86_64-cpython-313/functorch/einops copying functorch/einops/_parsing.py -> build/lib.linux-x86_64-cpython-313/functorch/einops copying functorch/einops/rearrange.py -> build/lib.linux-x86_64-cpython-313/functorch/einops copying functorch/einops/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/einops creating build/lib.linux-x86_64-cpython-313/functorch/experimental copying functorch/experimental/control_flow.py -> build/lib.linux-x86_64-cpython-313/functorch/experimental copying functorch/experimental/ops.py -> build/lib.linux-x86_64-cpython-313/functorch/experimental copying functorch/experimental/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/experimental creating build/lib.linux-x86_64-cpython-313/functorch/compile copying functorch/compile/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/compile creating build/lib.linux-x86_64-cpython-313/functorch/_src copying functorch/_src/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/_src creating build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/batch_tensor.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/tree_map.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/delayed_mul_tensor.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/reference.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/magic_trace.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/wrap_type.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/op_properties.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/dim.py -> build/lib.linux-x86_64-cpython-313/functorch/dim copying functorch/dim/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/dim creating build/lib.linux-x86_64-cpython-313/functorch/_src/eager_transforms copying functorch/_src/eager_transforms/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/_src/eager_transforms creating build/lib.linux-x86_64-cpython-313/functorch/_src/make_functional copying functorch/_src/make_functional/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/_src/make_functional creating build/lib.linux-x86_64-cpython-313/functorch/_src/vmap copying functorch/_src/vmap/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/_src/vmap creating build/lib.linux-x86_64-cpython-313/functorch/_src/aot_autograd copying functorch/_src/aot_autograd/__init__.py -> build/lib.linux-x86_64-cpython-313/functorch/_src/aot_autograd creating build/lib.linux-x86_64-cpython-313/torchgen/executorch copying torchgen/executorch/parse.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch copying torchgen/executorch/model.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch copying torchgen/executorch/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch creating build/lib.linux-x86_64-cpython-313/torchgen/dest copying torchgen/dest/lazy_ir.py -> build/lib.linux-x86_64-cpython-313/torchgen/dest copying torchgen/dest/native_functions.py -> build/lib.linux-x86_64-cpython-313/torchgen/dest copying torchgen/dest/register_dispatch_key.py -> build/lib.linux-x86_64-cpython-313/torchgen/dest copying torchgen/dest/lazy_ts_lowering.py -> build/lib.linux-x86_64-cpython-313/torchgen/dest copying torchgen/dest/ufunc.py -> build/lib.linux-x86_64-cpython-313/torchgen/dest copying torchgen/dest/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/dest creating build/lib.linux-x86_64-cpython-313/torchgen/selective_build copying torchgen/selective_build/operator.py -> build/lib.linux-x86_64-cpython-313/torchgen/selective_build copying torchgen/selective_build/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/selective_build copying torchgen/selective_build/selector.py -> build/lib.linux-x86_64-cpython-313/torchgen/selective_build creating build/lib.linux-x86_64-cpython-313/torchgen/aoti copying torchgen/aoti/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/aoti copying torchgen/aoti/fallback_ops.py -> build/lib.linux-x86_64-cpython-313/torchgen/aoti creating build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/cpp.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/functionalization.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/python.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/unboxing.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/translate.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/lazy.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/structured.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/autograd.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/dispatcher.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/ufunc.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/native.py -> build/lib.linux-x86_64-cpython-313/torchgen/api copying torchgen/api/meta.py -> build/lib.linux-x86_64-cpython-313/torchgen/api creating build/lib.linux-x86_64-cpython-313/torchgen/operator_versions copying torchgen/operator_versions/gen_mobile_upgraders_constant.py -> build/lib.linux-x86_64-cpython-313/torchgen/operator_versions copying torchgen/operator_versions/gen_mobile_upgraders.py -> build/lib.linux-x86_64-cpython-313/torchgen/operator_versions copying torchgen/operator_versions/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/operator_versions creating build/lib.linux-x86_64-cpython-313/torchgen/static_runtime copying torchgen/static_runtime/gen_static_runtime_ops.py -> build/lib.linux-x86_64-cpython-313/torchgen/static_runtime copying torchgen/static_runtime/config.py -> build/lib.linux-x86_64-cpython-313/torchgen/static_runtime copying torchgen/static_runtime/generator.py -> build/lib.linux-x86_64-cpython-313/torchgen/static_runtime copying torchgen/static_runtime/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/static_runtime creating build/lib.linux-x86_64-cpython-313/torchgen/executorch/api copying torchgen/executorch/api/et_cpp.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api copying torchgen/executorch/api/unboxing.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api copying torchgen/executorch/api/custom_ops.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api copying torchgen/executorch/api/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api creating build/lib.linux-x86_64-cpython-313/torchgen/executorch/api/types copying torchgen/executorch/api/types/signatures.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api/types copying torchgen/executorch/api/types/types.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api/types copying torchgen/executorch/api/types/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/executorch/api/types creating build/lib.linux-x86_64-cpython-313/torchgen/api/types copying torchgen/api/types/signatures.py -> build/lib.linux-x86_64-cpython-313/torchgen/api/types copying torchgen/api/types/types_base.py -> build/lib.linux-x86_64-cpython-313/torchgen/api/types copying torchgen/api/types/types.py -> build/lib.linux-x86_64-cpython-313/torchgen/api/types copying torchgen/api/types/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/api/types creating build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/random.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/jiterator.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/memory.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/_memory_viz.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/_sanitizer.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/nvtx.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/_gpu_trace.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/nccl.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/graphs.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/sparse.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/profiler.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/streams.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/comm.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/tunable.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/error.py -> build/lib.linux-x86_64-cpython-313/torch/cuda copying torch/cuda/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/cuda creating build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_state.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_ir_utils.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_pickle.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_decomposition_utils.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/annotations.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_script.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_freeze.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_recursive.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_trace.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_fuser.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/unsupported_tensor_ops.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_await.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/quantized.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/frontend.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_serialization.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_decompositions.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_async.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_builtins.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/generate_bytecode.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_shape_functions.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_dataclass_impls.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/supported_ops.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_logging.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_check.py -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/jit/_monkeytype_config.py -> build/lib.linux-x86_64-cpython-313/torch/jit creating build/lib.linux-x86_64-cpython-313/torch/mtia copying torch/mtia/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/mtia copying torch/mtia/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/mtia creating build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/extract_compiled_graph.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/config.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/device_context.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/ir_cache.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/metrics.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/computation.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/ts_backend.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/closure.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/tensor_factory_functions.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/debug.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy copying torch/_lazy/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_lazy creating build/lib.linux-x86_64-cpython-313/torch/special copying torch/special/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/special creating build/lib.linux-x86_64-cpython-313/torch/cpu copying torch/cpu/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/cpu creating build/lib.linux-x86_64-cpython-313/torch/ao copying torch/ao/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao creating build/lib.linux-x86_64-cpython-313/torch/contrib copying torch/contrib/_tensorboard_vis.py -> build/lib.linux-x86_64-cpython-313/torch/contrib copying torch/contrib/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/contrib creating build/lib.linux-x86_64-cpython-313/torch/testing copying torch/testing/_comparison.py -> build/lib.linux-x86_64-cpython-313/torch/testing copying torch/testing/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing copying torch/testing/_creation.py -> build/lib.linux-x86_64-cpython-313/torch/testing copying torch/testing/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing creating build/lib.linux-x86_64-cpython-313/torch/nested copying torch/nested/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nested creating build/lib.linux-x86_64-cpython-313/torch/backends copying torch/backends/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends creating build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/forward_ad.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/profiler_legacy.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/gradcheck.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/grad_mode.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/profiler_util.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/anomaly_mode.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/function.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/profiler.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/variable.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/functional.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/graph.py -> build/lib.linux-x86_64-cpython-313/torch/autograd copying torch/autograd/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/autograd creating build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/abstract_impl.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/simple_registry.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/autograd.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/infer_schema.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/custom_ops.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/fake_class_registry.py -> build/lib.linux-x86_64-cpython-313/torch/_library copying torch/_library/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_library creating build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/asgd.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/lbfgs.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/rprop.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/adamw.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/adagrad.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/adamax.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/radam.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/adadelta.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/sgd.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/sparse_adam.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/rmsprop.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/lr_scheduler.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/swa_utils.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/nadam.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/_functional.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/optim copying torch/optim/adam.py -> build/lib.linux-x86_64-cpython-313/torch/optim creating build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/unflatten.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/_remove_effect_tokens_pass.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/_trace.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/_tree_utils.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/exported_program.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/graph_signature.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/_remove_auto_functionalized_pass.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/custom_obj.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/_unlift.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/_safeguard.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/dynamic_shapes.py -> build/lib.linux-x86_64-cpython-313/torch/export copying torch/export/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/export creating build/lib.linux-x86_64-cpython-313/torch/amp copying torch/amp/grad_scaler.py -> build/lib.linux-x86_64-cpython-313/torch/amp copying torch/amp/autocast_mode.py -> build/lib.linux-x86_64-cpython-313/torch/amp copying torch/amp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/amp creating build/lib.linux-x86_64-cpython-313/torch/func copying torch/func/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/func creating build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/symbolic_convert.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/funcname_cache.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/guards.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/resume_execution.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/external_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/codegen.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/device_interface.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/debug_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/config.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/cache_size.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/logging.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/test_case.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/source.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/comptime.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/convert_frame.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/types.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/tensor_version_op.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/_trace_wrapped_higher_order_op.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/test_minifier_common.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/eval_frame.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/code_context.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/compiled_autograd.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/profiler.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/output_graph.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/exc.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/mutation_guard.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/create_parameter_op.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/replay_record.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/decorators.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/side_effects.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/testing.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/trace_rules.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/polyfill.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/current_scope_id.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/callback.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/hooks.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/bytecode_analysis.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo copying torch/_dynamo/bytecode_transformation.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo creating build/lib.linux-x86_64-cpython-313/torch/compiler copying torch/compiler/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/compiler creating build/lib.linux-x86_64-cpython-313/torch/_decomp copying torch/_decomp/decompositions_for_rng.py -> build/lib.linux-x86_64-cpython-313/torch/_decomp copying torch/_decomp/decompositions.py -> build/lib.linux-x86_64-cpython-313/torch/_decomp copying torch/_decomp/decompositions_for_jvp.py -> build/lib.linux-x86_64-cpython-313/torch/_decomp copying torch/_decomp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_decomp creating build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/wrap.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/auto_functionalize.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/associative_scan.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/map.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/effects.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/torchbind.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/strict_mode.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/triton_kernel_wrap.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/cond.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/while_loop.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/flex_attention.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/out_dtype.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops copying torch/_higher_order_ops/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_higher_order_ops creating build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/importer.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/find_file_dependencies.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_mangling.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/file_structure_representation.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/package_importer.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_stdlib.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_importlib.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_package_pickler.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_digraph.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_mock.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_package_unpickler.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/_directory_reader.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/package_exporter.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/package copying torch/package/glob_group.py -> build/lib.linux-x86_64-cpython-313/torch/package creating build/lib.linux-x86_64-cpython-313/torch/masked copying torch/masked/_ops.py -> build/lib.linux-x86_64-cpython-313/torch/masked copying torch/masked/_docs.py -> build/lib.linux-x86_64-cpython-313/torch/masked copying torch/masked/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/masked creating build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/constraint_registry.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/kumaraswamy.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/distribution.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/multinomial.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/gamma.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/transforms.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/log_normal.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/normal.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/poisson.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/cauchy.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/chi2.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/binomial.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/mixture_same_family.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/lowrank_multivariate_normal.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/continuous_bernoulli.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/exp_family.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/gumbel.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/kl.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/lkj_cholesky.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/half_normal.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/geometric.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/multivariate_normal.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/weibull.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/von_mises.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/fishersnedecor.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/pareto.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/inverse_gamma.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/independent.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/dirichlet.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/beta.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/one_hot_categorical.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/negative_binomial.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/laplace.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/bernoulli.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/uniform.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/constraints.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/exponential.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/relaxed_categorical.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/studentT.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/logistic_normal.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/wishart.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/relaxed_bernoulli.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/categorical.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/transformed_distribution.py -> build/lib.linux-x86_64-cpython-313/torch/distributions copying torch/distributions/half_cauchy.py -> build/lib.linux-x86_64-cpython-313/torch/distributions creating build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/fuse_modules.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/quantization_mappings.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/utils.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/qconfig.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/stubs.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/quant_type.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/_quantized_conversions.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/quantize_fx.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/quantize.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/_numeric_suite.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/fuser_method_mappings.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/_numeric_suite_fx.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/quantize_jit.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/observer.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/quantization copying torch/quantization/fake_quantize.py -> build/lib.linux-x86_64-cpython-313/torch/quantization creating build/lib.linux-x86_64-cpython-313/torch/mps copying torch/mps/event.py -> build/lib.linux-x86_64-cpython-313/torch/mps copying torch/mps/profiler.py -> build/lib.linux-x86_64-cpython-313/torch/mps copying torch/mps/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/mps creating build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/functional_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/fake_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/schema_check_mode.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/fake_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/meta_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/fake_impls.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses copying torch/_subclasses/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_subclasses creating build/lib.linux-x86_64-cpython-313/torch/monitor copying torch/monitor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/monitor creating build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/mkldnn_lowerings.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/cudagraph_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/virtualized.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/cudagraph_trees.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/comm_analysis.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/config.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/ir.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/index_propagation.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/freezing.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/ops_handler.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/dependencies.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/cpp_builder.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/test_case.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/scheduler.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/comms.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/codecache.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/autotune_process.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/select_algorithm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/decomposition.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/bounds.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/optimize_indexing.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/metrics.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/quantized_lowerings.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/sizevars.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/wrapper_benchmark.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/subgraph_lowering.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/pattern_matcher.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/test_operators.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/remote_cache.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/exc.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/async_compile.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/debug.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/constant_folding.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/hooks.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/lowering.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/fx_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/graph.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/compile_fx.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor copying torch/_inductor/inductor_prims.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor creating build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/mobile_optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_content_store.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/collect_env.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_import_utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_device.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_freeze.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/module_tracker.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_mode_utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/checkpoint.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/file_baton.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/throughput_benchmark.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/cpp_extension.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_cxx_pytree.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_typing_utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_contextlib.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_get_clean_triton.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_python_dispatch.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/flop_counter.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_pytree.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/backend_registration.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_config_module.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_exposed_in.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/cpp_backtrace.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/show_pickle.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_stats.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_traceback.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_triton.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/bundled_inputs.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/model_zoo.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/deterministic.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/mkldnn.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_foreach_utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_zip.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/hooks.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/_cpp_extension_versioner.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/weak.py -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/utils/dlpack.py -> build/lib.linux-x86_64-cpython-313/torch/utils creating build/lib.linux-x86_64-cpython-313/torch/_dispatch copying torch/_dispatch/python.py -> build/lib.linux-x86_64-cpython-313/torch/_dispatch copying torch/_dispatch/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_dispatch creating build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/_pattern_matcher.py -> build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/_memory_profiler.py -> build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/profiler.py -> build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/python_tracer.py -> build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/itt.py -> build/lib.linux-x86_64-cpython-313/torch/profiler copying torch/profiler/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/profiler creating build/lib.linux-x86_64-cpython-313/torch/_prims copying torch/_prims/rng_prims.py -> build/lib.linux-x86_64-cpython-313/torch/_prims copying torch/_prims/context.py -> build/lib.linux-x86_64-cpython-313/torch/_prims copying torch/_prims/executor.py -> build/lib.linux-x86_64-cpython-313/torch/_prims copying torch/_prims/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_prims copying torch/_prims/debug_prims.py -> build/lib.linux-x86_64-cpython-313/torch/_prims creating build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/random.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_dtypes.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_ndarray.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_funcs.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_getlimits.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_casting_dicts.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_binary_ufuncs_impl.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/fft.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_dtypes_impl.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_funcs_impl.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_unary_ufuncs_impl.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_normalizations.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/linalg.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_ufuncs.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_reductions_impl.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy copying torch/_numpy/_util.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy creating build/lib.linux-x86_64-cpython-313/torch/_vendor copying torch/_vendor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_vendor creating build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/_functional_collectives_impl.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/run.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/rendezvous.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/constants.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/device_mesh.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/distributed_c10d.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/_functional_collectives.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/_composable_state.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/c10d_logger.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/logging_handlers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/remote_device.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/launch.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/argparse_util.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/collective_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed copying torch/distributed/_state_dict_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed creating build/lib.linux-x86_64-cpython-313/torch/_prims_common copying torch/_prims_common/wrappers.py -> build/lib.linux-x86_64-cpython-313/torch/_prims_common copying torch/_prims_common/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_prims_common creating build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/_lazy_graph_module.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/_symbolic_trace.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/_compatibility.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/config.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/traceback.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/subgraph_rewriter.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/node.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/operator_schemas.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/_pytree.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/proxy.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/interpreter.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/graph_module.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/tensor_type.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/annotate.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/graph.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/fx/immutable_collections.py -> build/lib.linux-x86_64-cpython-313/torch/fx creating build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/deprecated.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/compile_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/aot_autograd.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/benchmark_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/batch_norm_replacement.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/config.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/functional_call.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/partitioners.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/top_operators_github_usage.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/apis.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/pyfunctorch.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/fx_minifier.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/make_functional.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/pytree_hacks.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/python_key.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/compilers.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/eager_transforms.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/autograd_function.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch copying torch/_functorch/vmap.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch creating build/lib.linux-x86_64-cpython-313/torch/_custom_op copying torch/_custom_op/impl.py -> build/lib.linux-x86_64-cpython-313/torch/_custom_op copying torch/_custom_op/autograd.py -> build/lib.linux-x86_64-cpython-313/torch/_custom_op copying torch/_custom_op/functional.py -> build/lib.linux-x86_64-cpython-313/torch/_custom_op copying torch/_custom_op/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_custom_op creating build/lib.linux-x86_64-cpython-313/torch/multiprocessing copying torch/multiprocessing/queue.py -> build/lib.linux-x86_64-cpython-313/torch/multiprocessing copying torch/multiprocessing/spawn.py -> build/lib.linux-x86_64-cpython-313/torch/multiprocessing copying torch/multiprocessing/reductions.py -> build/lib.linux-x86_64-cpython-313/torch/multiprocessing copying torch/multiprocessing/pool.py -> build/lib.linux-x86_64-cpython-313/torch/multiprocessing copying torch/multiprocessing/_atfork.py -> build/lib.linux-x86_64-cpython-313/torch/multiprocessing copying torch/multiprocessing/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/multiprocessing creating build/lib.linux-x86_64-cpython-313/torch/_refs copying torch/_refs/fft.py -> build/lib.linux-x86_64-cpython-313/torch/_refs copying torch/_refs/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_refs copying torch/_refs/_conversions.py -> build/lib.linux-x86_64-cpython-313/torch/_refs creating build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset18.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/errors.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset19.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_exporter_states.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/verification.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset20.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset10.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset14.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_helper.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset17.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset11.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset8.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset9.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_onnx_supported_ops.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_experimental.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_globals.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_caffe2.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset12.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset13.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_constants.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/operators.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset16.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset7.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_type_utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/symbolic_opset15.py -> build/lib.linux-x86_64-cpython-313/torch/onnx copying torch/onnx/_deprecation.py -> build/lib.linux-x86_64-cpython-313/torch/onnx creating build/lib.linux-x86_64-cpython-313/torch/linalg copying torch/linalg/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/linalg creating build/lib.linux-x86_64-cpython-313/torch/signal copying torch/signal/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/signal creating build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/cpp.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/grad.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/common_types.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/_reduction.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/init.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/functional.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/parameter.py -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn creating build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/non_strict_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/converter.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/wrappers.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/verifier.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/exported_program.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/error.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/pass_base.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_export copying torch/_export/tools.py -> build/lib.linux-x86_64-cpython-313/torch/_export creating build/lib.linux-x86_64-cpython-313/torch/futures copying torch/futures/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/futures creating build/lib.linux-x86_64-cpython-313/torch/sparse copying torch/sparse/_semi_structured_ops.py -> build/lib.linux-x86_64-cpython-313/torch/sparse copying torch/sparse/_semi_structured_conversions.py -> build/lib.linux-x86_64-cpython-313/torch/sparse copying torch/sparse/_triton_ops_meta.py -> build/lib.linux-x86_64-cpython-313/torch/sparse copying torch/sparse/_triton_ops.py -> build/lib.linux-x86_64-cpython-313/torch/sparse copying torch/sparse/semi_structured.py -> build/lib.linux-x86_64-cpython-313/torch/sparse copying torch/sparse/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/sparse creating build/lib.linux-x86_64-cpython-313/torch/_strobelight copying torch/_strobelight/cli_function_profiler.py -> build/lib.linux-x86_64-cpython-313/torch/_strobelight copying torch/_strobelight/compile_time_profiler.py -> build/lib.linux-x86_64-cpython-313/torch/_strobelight copying torch/_strobelight/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_strobelight creating build/lib.linux-x86_64-cpython-313/torch/_awaits copying torch/_awaits/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_awaits creating build/lib.linux-x86_64-cpython-313/torch/xpu copying torch/xpu/random.py -> build/lib.linux-x86_64-cpython-313/torch/xpu copying torch/xpu/_gpu_trace.py -> build/lib.linux-x86_64-cpython-313/torch/xpu copying torch/xpu/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/xpu copying torch/xpu/streams.py -> build/lib.linux-x86_64-cpython-313/torch/xpu copying torch/xpu/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/xpu creating build/lib.linux-x86_64-cpython-313/torch/_logging copying torch/_logging/_registrations.py -> build/lib.linux-x86_64-cpython-313/torch/_logging copying torch/_logging/structured.py -> build/lib.linux-x86_64-cpython-313/torch/_logging copying torch/_logging/_internal.py -> build/lib.linux-x86_64-cpython-313/torch/_logging copying torch/_logging/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_logging creating build/lib.linux-x86_64-cpython-313/torch/fft copying torch/fft/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fft creating build/lib.linux-x86_64-cpython-313/torch/cuda/amp copying torch/cuda/amp/common.py -> build/lib.linux-x86_64-cpython-313/torch/cuda/amp copying torch/cuda/amp/grad_scaler.py -> build/lib.linux-x86_64-cpython-313/torch/cuda/amp copying torch/cuda/amp/autocast_mode.py -> build/lib.linux-x86_64-cpython-313/torch/cuda/amp copying torch/cuda/amp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/cuda/amp creating build/lib.linux-x86_64-cpython-313/torch/jit/_passes copying torch/jit/_passes/_property_propagation.py -> build/lib.linux-x86_64-cpython-313/torch/jit/_passes copying torch/jit/_passes/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/jit/_passes creating build/lib.linux-x86_64-cpython-313/torch/jit/mobile copying torch/jit/mobile/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/jit/mobile creating build/lib.linux-x86_64-cpython-313/torch/cpu/amp copying torch/cpu/amp/grad_scaler.py -> build/lib.linux-x86_64-cpython-313/torch/cpu/amp copying torch/cpu/amp/autocast_mode.py -> build/lib.linux-x86_64-cpython-313/torch/cpu/amp copying torch/cpu/amp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/cpu/amp creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning copying torch/ao/pruning/_mappings.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning copying torch/ao/pruning/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning creating build/lib.linux-x86_64-cpython-313/torch/ao/ns copying torch/ao/ns/_numeric_suite.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns copying torch/ao/ns/_numeric_suite_fx.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns copying torch/ao/ns/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/fuse_modules.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/quantization_mappings.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/qconfig.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/stubs.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/quant_type.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/quantize_fx.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/quantize_pt2e.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/quantize.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/fuser_method_mappings.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/_learnable_fake_quantize.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/quantize_jit.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/observer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/_equalize.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/_correct_bias.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/qconfig_mapping.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization copying torch/ao/quantization/fake_quantize.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization creating build/lib.linux-x86_64-cpython-313/torch/ao/nn copying torch/ao/nn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental copying torch/ao/pruning/_experimental/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/sparsifier copying torch/ao/pruning/sparsifier/base_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/sparsifier copying torch/ao/pruning/sparsifier/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/sparsifier copying torch/ao/pruning/sparsifier/nearly_diagonal_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/sparsifier copying torch/ao/pruning/sparsifier/weight_norm_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/sparsifier copying torch/ao/pruning/sparsifier/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/sparsifier creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/scheduler copying torch/ao/pruning/scheduler/lambda_scheduler.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/scheduler copying torch/ao/pruning/scheduler/base_scheduler.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/scheduler copying torch/ao/pruning/scheduler/cubic_scheduler.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/scheduler copying torch/ao/pruning/scheduler/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/scheduler creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier copying torch/ao/pruning/_experimental/data_sparsifier/data_norm_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier copying torch/ao/pruning/_experimental/data_sparsifier/base_data_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier copying torch/ao/pruning/_experimental/data_sparsifier/quantization_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier copying torch/ao/pruning/_experimental/data_sparsifier/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/lstm_saliency_pruner.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/saliency_pruner.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/FPGM_pruner.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/match_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/prune_functions.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/parametrization.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/base_structured_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner copying torch/ao/pruning/_experimental/pruner/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/pruner creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/activation_sparsifier copying torch/ao/pruning/_experimental/activation_sparsifier/activation_sparsifier.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/activation_sparsifier copying torch/ao/pruning/_experimental/activation_sparsifier/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/activation_sparsifier creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_scheduler copying torch/ao/pruning/_experimental/data_scheduler/base_data_scheduler.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_scheduler copying torch/ao/pruning/_experimental/data_scheduler/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_scheduler creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier/lightning copying torch/ao/pruning/_experimental/data_sparsifier/lightning/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier/lightning creating build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks copying torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks/_data_sparstity_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks copying torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks/data_sparsity.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks copying torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/pruning/_experimental/data_sparsifier/lightning/callbacks creating build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/graph_passes.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/ns_types.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/pattern_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/mappings.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/qconfig_multi_mapping.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/n_shadows_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/weight_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/graph_matcher.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx copying torch/ao/ns/fx/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/ns/fx creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/tensorrt.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/qnnpack.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/onednn.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/observation_type.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/_common_operator_config_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/backend_config.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/x86.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/_qnnpack_pt2e.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/fbgemm.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/native.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config copying torch/ao/quantization/backend_config/executorch.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/backend_config creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/x86_inductor_quantizer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/xnnpack_quantizer_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/composable_quantizer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/embedding_quantizer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/xnnpack_quantizer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/quantizer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer copying torch/ao/quantization/quantizer/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/quantizer creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/convert.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/tracer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/pattern_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/lower_to_fbgemm.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/lstm_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/qconfig_mapping_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/custom_config.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/match_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/quantize_handler.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/lower_to_qnnpack.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/fuse_handler.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/_decomposed.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/graph_module.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/prepare.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/fuse.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/_lower_to_native_backend.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/_equalize.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx copying torch/ao/quantization/fx/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/generate_numeric_debug_handle.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/graph_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/port_metadata_pass.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/prepare.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/export_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/duplicate_dq_pass.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/qat_utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e copying torch/ao/quantization/pt2e/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx/_model_report copying torch/ao/quantization/fx/_model_report/detector.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx/_model_report copying torch/ao/quantization/fx/_model_report/model_report_visualizer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx/_model_report copying torch/ao/quantization/fx/_model_report/model_report_observer.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx/_model_report copying torch/ao/quantization/fx/_model_report/model_report.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx/_model_report copying torch/ao/quantization/fx/_model_report/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/fx/_model_report creating build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e/representation copying torch/ao/quantization/pt2e/representation/rewrite.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e/representation copying torch/ao/quantization/pt2e/representation/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/quantization/pt2e/representation creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic copying torch/ao/nn/intrinsic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized copying torch/ao/nn/quantized/functional.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized copying torch/ao/nn/quantized/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat copying torch/ao/nn/qat/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse copying torch/ao/nn/sparse/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantizable copying torch/ao/nn/quantizable/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantizable creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized copying torch/ao/nn/intrinsic/quantized/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/modules copying torch/ao/nn/intrinsic/modules/fused.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/modules copying torch/ao/nn/intrinsic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat copying torch/ao/nn/intrinsic/qat/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/modules copying torch/ao/nn/intrinsic/quantized/modules/conv_relu.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/modules copying torch/ao/nn/intrinsic/quantized/modules/bn_relu.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/modules copying torch/ao/nn/intrinsic/quantized/modules/conv_add.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/modules copying torch/ao/nn/intrinsic/quantized/modules/linear_relu.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/modules copying torch/ao/nn/intrinsic/quantized/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/dynamic copying torch/ao/nn/intrinsic/quantized/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/dynamic creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/dynamic/modules copying torch/ao/nn/intrinsic/quantized/dynamic/modules/linear_relu.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/dynamic/modules copying torch/ao/nn/intrinsic/quantized/dynamic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/quantized/dynamic/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat/modules copying torch/ao/nn/intrinsic/qat/modules/linear_fused.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat/modules copying torch/ao/nn/intrinsic/qat/modules/linear_relu.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat/modules copying torch/ao/nn/intrinsic/qat/modules/conv_fused.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat/modules copying torch/ao/nn/intrinsic/qat/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/intrinsic/qat/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference copying torch/ao/nn/quantized/reference/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/dropout.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/normalization.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/embedding_ops.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/activation.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/batchnorm.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/functional_modules.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules copying torch/ao/nn/quantized/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic copying torch/ao/nn/quantized/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules copying torch/ao/nn/quantized/reference/modules/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules copying torch/ao/nn/quantized/reference/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules copying torch/ao/nn/quantized/reference/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules copying torch/ao/nn/quantized/reference/modules/sparse.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules copying torch/ao/nn/quantized/reference/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules copying torch/ao/nn/quantized/reference/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/reference/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic/modules copying torch/ao/nn/quantized/dynamic/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic/modules copying torch/ao/nn/quantized/dynamic/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic/modules copying torch/ao/nn/quantized/dynamic/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic/modules copying torch/ao/nn/quantized/dynamic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantized/dynamic/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/modules copying torch/ao/nn/qat/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/modules copying torch/ao/nn/qat/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/modules copying torch/ao/nn/qat/modules/embedding_ops.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/modules copying torch/ao/nn/qat/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/dynamic copying torch/ao/nn/qat/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/dynamic creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/dynamic/modules copying torch/ao/nn/qat/dynamic/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/dynamic/modules copying torch/ao/nn/qat/dynamic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/qat/dynamic/modules creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized copying torch/ao/nn/sparse/quantized/utils.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized copying torch/ao/nn/sparse/quantized/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized copying torch/ao/nn/sparse/quantized/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized/dynamic copying torch/ao/nn/sparse/quantized/dynamic/linear.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized/dynamic copying torch/ao/nn/sparse/quantized/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/sparse/quantized/dynamic creating build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantizable/modules copying torch/ao/nn/quantizable/modules/activation.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantizable/modules copying torch/ao/nn/quantizable/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantizable/modules copying torch/ao/nn/quantizable/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/ao/nn/quantizable/modules creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/dist_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_pruning.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/hypothesis_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_methods_invocations.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/jit_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_dist_composable.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_mkldnn.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/logging_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_modules.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_subclass.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/composite_compliance.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/quantization_torch_package_models.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_distributed.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_quantized.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/torchbind_impls.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_quantization.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_dtype.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/inductor_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/custom_op_db.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_optimizers.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_cuda.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/autograd_function_db.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/two_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_nn.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/hop_db.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/dynamo_test_failures.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/check_kernel_launches.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/logging_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/static_module.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_jit.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/autocast_test_lists.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/jit_metaprogramming_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/triton_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_fsdp.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal copying torch/testing/_internal/common_device_type.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/test_module copying torch/testing/_internal/test_module/future_div.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/test_module copying torch/testing/_internal/test_module/no_future_div.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/test_module copying torch/testing/_internal/test_module/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/test_module creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/codegen copying torch/testing/_internal/codegen/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/codegen creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo copying torch/testing/_internal/opinfo/core.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo copying torch/testing/_internal/opinfo/utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo copying torch/testing/_internal/opinfo/refs.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo copying torch/testing/_internal/opinfo/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests copying torch/testing/_internal/optests/aot_autograd.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests copying torch/testing/_internal/optests/autograd_registration.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests copying torch/testing/_internal/optests/fake_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests copying torch/testing/_internal/optests/make_fx.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests copying torch/testing/_internal/optests/generate_tests.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests copying torch/testing/_internal/optests/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/optests creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/rpc_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/ddp_under_dist_autograd_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/fake_pg.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/multi_threaded_pg.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/checkpoint_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/common_state_dict.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/distributed_utils.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/distributed_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed copying torch/testing/_internal/distributed/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/generated copying torch/testing/_internal/generated/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/generated copying torch/testing/_internal/generated/annotated_fn_args.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/generated creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/data copying torch/testing/_internal/data/network1.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/data copying torch/testing/_internal/data/network2.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/data copying torch/testing/_internal/data/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/data creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/special.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/_masked.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/signal.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/fft.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/sparse.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/linalg.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions copying torch/testing/_internal/opinfo/definitions/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/opinfo/definitions creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard copying torch/testing/_internal/distributed/_shard/test_common.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard copying torch/testing/_internal/distributed/_shard/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/rpc_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/dist_optimizer_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/faulty_agent_rpc_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/dist_autograd_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/rpc_agent_test_fixture.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/tensorpipe_rpc_agent_test_fixture.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc copying torch/testing/_internal/distributed/rpc/faulty_rpc_agent_test_fixture.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_tensor copying torch/testing/_internal/distributed/_tensor/common_dtensor.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_tensor copying torch/testing/_internal/distributed/_tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_tensor creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/nn copying torch/testing/_internal/distributed/nn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/nn creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard/sharded_tensor copying torch/testing/_internal/distributed/_shard/sharded_tensor/_test_st_common.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard/sharded_tensor copying torch/testing/_internal/distributed/_shard/sharded_tensor/_test_ops_common.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard/sharded_tensor copying torch/testing/_internal/distributed/_shard/sharded_tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/_shard/sharded_tensor creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/jit copying torch/testing/_internal/distributed/rpc/jit/rpc_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/jit copying torch/testing/_internal/distributed/rpc/jit/rpc_test_faulty.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/jit copying torch/testing/_internal/distributed/rpc/jit/dist_autograd_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/jit copying torch/testing/_internal/distributed/rpc/jit/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/jit creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/examples copying torch/testing/_internal/distributed/rpc/examples/reinforcement_learning_rpc_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/examples copying torch/testing/_internal/distributed/rpc/examples/parameter_server_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/examples copying torch/testing/_internal/distributed/rpc/examples/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/rpc/examples creating build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/nn/api copying torch/testing/_internal/distributed/nn/api/remote_module_test.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/nn/api copying torch/testing/_internal/distributed/nn/api/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/testing/_internal/distributed/nn/api creating build/lib.linux-x86_64-cpython-313/torch/nested/_internal copying torch/nested/_internal/sdpa.py -> build/lib.linux-x86_64-cpython-313/torch/nested/_internal copying torch/nested/_internal/nested_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/nested/_internal copying torch/nested/_internal/ops.py -> build/lib.linux-x86_64-cpython-313/torch/nested/_internal copying torch/nested/_internal/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nested/_internal creating build/lib.linux-x86_64-cpython-313/torch/backends/cuda copying torch/backends/cuda/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/cuda creating build/lib.linux-x86_64-cpython-313/torch/backends/xnnpack copying torch/backends/xnnpack/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/xnnpack creating build/lib.linux-x86_64-cpython-313/torch/backends/mha copying torch/backends/mha/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/mha creating build/lib.linux-x86_64-cpython-313/torch/backends/cpu copying torch/backends/cpu/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/cpu creating build/lib.linux-x86_64-cpython-313/torch/backends/cudnn copying torch/backends/cudnn/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/backends/cudnn copying torch/backends/cudnn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/cudnn creating build/lib.linux-x86_64-cpython-313/torch/backends/opt_einsum copying torch/backends/opt_einsum/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/opt_einsum creating build/lib.linux-x86_64-cpython-313/torch/backends/mkldnn copying torch/backends/mkldnn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/mkldnn creating build/lib.linux-x86_64-cpython-313/torch/backends/_coreml copying torch/backends/_coreml/preprocess.py -> build/lib.linux-x86_64-cpython-313/torch/backends/_coreml copying torch/backends/_coreml/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/_coreml creating build/lib.linux-x86_64-cpython-313/torch/backends/quantized copying torch/backends/quantized/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/quantized creating build/lib.linux-x86_64-cpython-313/torch/backends/mps copying torch/backends/mps/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/mps creating build/lib.linux-x86_64-cpython-313/torch/backends/nnpack copying torch/backends/nnpack/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/nnpack creating build/lib.linux-x86_64-cpython-313/torch/backends/_nnapi copying torch/backends/_nnapi/serializer.py -> build/lib.linux-x86_64-cpython-313/torch/backends/_nnapi copying torch/backends/_nnapi/prepare.py -> build/lib.linux-x86_64-cpython-313/torch/backends/_nnapi copying torch/backends/_nnapi/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/_nnapi creating build/lib.linux-x86_64-cpython-313/torch/backends/mkl copying torch/backends/mkl/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/mkl creating build/lib.linux-x86_64-cpython-313/torch/backends/xeon copying torch/backends/xeon/run_cpu.py -> build/lib.linux-x86_64-cpython-313/torch/backends/xeon copying torch/backends/xeon/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/xeon creating build/lib.linux-x86_64-cpython-313/torch/backends/openmp copying torch/backends/openmp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/backends/openmp creating build/lib.linux-x86_64-cpython-313/torch/autograd/_functions copying torch/autograd/_functions/utils.py -> build/lib.linux-x86_64-cpython-313/torch/autograd/_functions copying torch/autograd/_functions/tensor.py -> build/lib.linux-x86_64-cpython-313/torch/autograd/_functions copying torch/autograd/_functions/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/autograd/_functions creating build/lib.linux-x86_64-cpython-313/torch/optim/_multi_tensor copying torch/optim/_multi_tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/optim/_multi_tensor creating build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/ctx_manager.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/higher_order_ops.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/nn_module.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/user_defined.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/lists.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/torch.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/lazy.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/sdpa.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/base.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/constant.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/iter.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/misc.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/functions.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/torch_function.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/dicts.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/script_object.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/builder.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/tensor.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/builtin.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables copying torch/_dynamo/variables/distributed.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/variables creating build/lib.linux-x86_64-cpython-313/torch/_dynamo/repro copying torch/_dynamo/repro/after_dynamo.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/repro copying torch/_dynamo/repro/after_aot.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/repro copying torch/_dynamo/repro/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/repro creating build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/tensorrt.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/torchxla.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/common.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/onnxrt.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/debugging.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/inductor.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/cudagraphs.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/tvm.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/registry.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends copying torch/_dynamo/backends/distributed.py -> build/lib.linux-x86_64-cpython-313/torch/_dynamo/backends creating build/lib.linux-x86_64-cpython-313/torch/package/analyze copying torch/package/analyze/find_first_use_of_broken_modules.py -> build/lib.linux-x86_64-cpython-313/torch/package/analyze copying torch/package/analyze/trace_dependencies.py -> build/lib.linux-x86_64-cpython-313/torch/package/analyze copying torch/package/analyze/is_from_package.py -> build/lib.linux-x86_64-cpython-313/torch/package/analyze copying torch/package/analyze/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/package/analyze creating build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/core.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/_ops_refs.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/binary.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/passthrough.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/creation.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/reductions.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/unary.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor copying torch/masked/maskedtensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/masked/maskedtensor creating build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/quantization_types.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/convert.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/utils.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/fusion_patterns.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/pattern_utils.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/quantization_patterns.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/match_utils.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/graph_module.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/prepare.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/fuse.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/_equalize.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx copying torch/quantization/fx/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/quantization/fx creating build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/dedupe_symint_uses.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/replace_random.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/joint_graph.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/numeric_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/pre_grad.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/misc_patterns.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/decompose_mem_bound_mm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/reinplace.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/fuse_attention.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/freezing_patterns.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/ddp_fusion.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/binary_folding.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/group_batch_fusion.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/mkldnn_fusion.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/quantization.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/split_cat.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/pad_mm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/micro_pipeline_tp.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/post_grad.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/efficient_conv_bn_eval.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes copying torch/_inductor/fx_passes/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes creating build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_gemm_template.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/triton_foreach.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cuda_combined_scheduling.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_template_kernel.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/multi_kernel.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/common.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/memory_planning.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/aoti_hipify_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/simd.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_micro_gemm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_wrapper_cuda.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/triton_split_scan.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_wrapper_cpu.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_template.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/wrapper.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/triton.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/cpp_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/triton_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/codegen_device_driver.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen copying torch/_inductor/codegen/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen creating build/lib.linux-x86_64-cpython-313/torch/_inductor/compile_worker copying torch/_inductor/compile_worker/watchdog.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/compile_worker copying torch/_inductor/compile_worker/__main__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/compile_worker copying torch/_inductor/compile_worker/subproc_pool.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/compile_worker copying torch/_inductor/compile_worker/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/compile_worker creating build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/runtime_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/triton_helpers.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/triton_heuristics.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/coordinate_descent_tuner.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/compile_tasks.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/hints.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime copying torch/_inductor/runtime/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/runtime creating build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/mm_common.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/conv.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/bmm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/unpack_mixed_mm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/mm_plus_mm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/flex_attention.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/mm.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel copying torch/_inductor/kernel/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/kernel creating build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_17.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_6.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_8.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_5.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_11.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/bmm_pattern.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_18.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/addmm_pattern.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_4.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_16.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_14.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_19.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_13.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_2.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_3.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_9.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_1.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/mm_pattern.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_15.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_12.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_10.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns copying torch/_inductor/fx_passes/serialized_patterns/_sfdp_pattern_7.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/fx_passes/serialized_patterns creating build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/cuda_env.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/gemm_template.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/cutlass_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/cuda_kernel.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/cutlass_epilogue_gen.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/cuda_cpp_scheduling.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/cuda_template.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda copying torch/_inductor/codegen/cuda/device_op_overrides.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda creating build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/xpu copying torch/_inductor/codegen/xpu/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/xpu copying torch/_inductor/codegen/xpu/device_op_overrides.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/xpu creating build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda/cutlass_lib_extensions copying torch/_inductor/codegen/cuda/cutlass_lib_extensions/gemm_operation_extensions.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda/cutlass_lib_extensions copying torch/_inductor/codegen/cuda/cutlass_lib_extensions/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/cuda/cutlass_lib_extensions creating build/lib.linux-x86_64-cpython-313/torch/utils/bottleneck copying torch/utils/bottleneck/__main__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/bottleneck copying torch/utils/bottleneck/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/bottleneck creating build/lib.linux-x86_64-cpython-313/torch/utils/jit copying torch/utils/jit/log_extract.py -> build/lib.linux-x86_64-cpython-313/torch/utils/jit copying torch/utils/jit/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/jit creating build/lib.linux-x86_64-cpython-313/torch/utils/benchmark copying torch/utils/benchmark/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark creating build/lib.linux-x86_64-cpython-313/torch/utils/backcompat copying torch/utils/backcompat/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/backcompat creating build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/singleton_int.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/reference.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/value_ranges.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/symbol.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/solve.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/functions.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/interp.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy copying torch/utils/_sympy/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_sympy creating build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/_embedding.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/_pytorch_graph.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/_onnx_graph.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/_proto_graph.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/_convert_np.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/summary.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard copying torch/utils/tensorboard/writer.py -> build/lib.linux-x86_64-cpython-313/torch/utils/tensorboard creating build/lib.linux-x86_64-cpython-313/torch/utils/hipify copying torch/utils/hipify/constants.py -> build/lib.linux-x86_64-cpython-313/torch/utils/hipify copying torch/utils/hipify/version.py -> build/lib.linux-x86_64-cpython-313/torch/utils/hipify copying torch/utils/hipify/cuda_to_hip_mappings.py -> build/lib.linux-x86_64-cpython-313/torch/utils/hipify copying torch/utils/hipify/hipify_python.py -> build/lib.linux-x86_64-cpython-313/torch/utils/hipify copying torch/utils/hipify/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/hipify creating build/lib.linux-x86_64-cpython-313/torch/utils/viz copying torch/utils/viz/_cycles.py -> build/lib.linux-x86_64-cpython-313/torch/utils/viz copying torch/utils/viz/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/viz creating build/lib.linux-x86_64-cpython-313/torch/utils/model_dump copying torch/utils/model_dump/__main__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/model_dump copying torch/utils/model_dump/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/model_dump creating build/lib.linux-x86_64-cpython-313/torch/utils/_strobelight copying torch/utils/_strobelight/cli_function_profiler.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_strobelight copying torch/utils/_strobelight/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/_strobelight creating build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/dataloader.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/backward_compatibility.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/dataset.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/graph_settings.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/sampler.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/graph.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data copying torch/utils/data/distributed.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data creating build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/compare.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/simple_timeit.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/spectral_ops_fuzz_test.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/fuzzer.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/op_benchmark.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/blas_compare_setup.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples copying torch/utils/benchmark/examples/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/examples creating build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/cpp_jit.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/sparse_fuzzer.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/compare.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/timer.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/common.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/fuzzer.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/_stubs.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/compile.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils creating build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers copying torch/utils/benchmark/op_fuzzers/sparse_unary.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers copying torch/utils/benchmark/op_fuzzers/sparse_binary.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers copying torch/utils/benchmark/op_fuzzers/binary.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers copying torch/utils/benchmark/op_fuzzers/unary.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers copying torch/utils/benchmark/op_fuzzers/spectral.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers copying torch/utils/benchmark/op_fuzzers/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/op_fuzzers creating build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper copying torch/utils/benchmark/utils/valgrind_wrapper/timer_interface.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper copying torch/utils/benchmark/utils/valgrind_wrapper/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper creating build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/utils/data/datapipes/datapipe.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/utils/data/datapipes/gen_pyi.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/utils/data/datapipes/_typing.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/utils/data/datapipes/_hook_iterator.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/utils/data/datapipes/_decorator.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/utils/data/datapipes/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes creating build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils copying torch/utils/data/_utils/fetch.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils copying torch/utils/data/_utils/pin_memory.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils copying torch/utils/data/_utils/signal_handling.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils copying torch/utils/data/_utils/worker.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils copying torch/utils/data/_utils/collate.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils copying torch/utils/data/_utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/_utils creating build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/routeddecoder.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/streamreader.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/combinatorics.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/filelister.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/combining.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/selecting.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/sharding.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/callable.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/fileopener.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter copying torch/utils/data/datapipes/iter/grouping.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/iter creating build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map copying torch/utils/data/datapipes/map/combinatorics.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map copying torch/utils/data/datapipes/map/utils.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map copying torch/utils/data/datapipes/map/combining.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map copying torch/utils/data/datapipes/map/callable.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map copying torch/utils/data/datapipes/map/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map copying torch/utils/data/datapipes/map/grouping.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/map creating build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/utils copying torch/utils/data/datapipes/utils/decoder.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/utils copying torch/utils/data/datapipes/utils/common.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/utils copying torch/utils/data/datapipes/utils/snapshot.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/utils copying torch/utils/data/datapipes/utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/utils creating build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/dataframe copying torch/utils/data/datapipes/dataframe/datapipes.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/dataframe copying torch/utils/data/datapipes/dataframe/structures.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/dataframe copying torch/utils/data/datapipes/dataframe/dataframes.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/dataframe copying torch/utils/data/datapipes/dataframe/dataframe_wrapper.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/dataframe copying torch/utils/data/datapipes/dataframe/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes/dataframe creating build/lib.linux-x86_64-cpython-313/torch/_numpy/testing copying torch/_numpy/testing/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy/testing copying torch/_numpy/testing/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_numpy/testing creating build/lib.linux-x86_64-cpython-313/torch/_vendor/packaging copying torch/_vendor/packaging/version.py -> build/lib.linux-x86_64-cpython-313/torch/_vendor/packaging copying torch/_vendor/packaging/_structures.py -> build/lib.linux-x86_64-cpython-313/torch/_vendor/packaging copying torch/_vendor/packaging/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_vendor/packaging creating build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/_unflatten.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/_backward.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/schedules.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/_IR.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/stage.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/microbatch.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining copying torch/distributed/pipelining/_debug.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/pipelining creating build/lib.linux-x86_64-cpython-313/torch/distributed/_sharded_tensor copying torch/distributed/_sharded_tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_sharded_tensor creating build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_optim_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_flat_param.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_wrap_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/sharded_grad_scaler.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/wrap.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_common_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_debug_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_unshard_param_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_init_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_traversal_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_limiter_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_fsdp_extensions.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_shard_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_runtime_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_exec_order_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/fully_sharded_data_parallel.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_dynamo_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_trace_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp copying torch/distributed/fsdp/_state_dict_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/fsdp creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/common_op_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/sharder.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/metadata.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/op_registry_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard copying torch/distributed/_shard/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard creating build/lib.linux-x86_64-cpython-313/torch/distributed/autograd copying torch/distributed/autograd/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/autograd creating build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/post_localSGD_optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_adadelta.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/apply_optimizer_in_backward.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_rmsprop.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_rprop.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_adamw.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/zero_redundancy_optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_adam.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/named_optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_adagrad.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_adamax.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/functional_sgd.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/distributed/optim/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim creating build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_traverse.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/planner.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_dedup_tensors.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/staging.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/optimizer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/storage.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/filesystem.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/resharding.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/stateful.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/format_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_checkpointer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/planner_helpers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/default_planner.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/logger.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_nested_dict.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/state_dict.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_storage_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/metadata.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/state_dict_saver.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_sharded_tensor_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/state_dict_loader.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/logging_handlers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_fsspec_filesystem.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/_dedup_save_plans.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint copying torch/distributed/checkpoint/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/checkpoint creating build/lib.linux-x86_64-cpython-313/torch/distributed/_cuda_p2p copying torch/distributed/_cuda_p2p/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_cuda_p2p creating build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/server_process_global_profiler.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/options.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/constants.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/backend_registry.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/internal.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/functions.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc copying torch/distributed/rpc/rref_proxy.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc creating build/lib.linux-x86_64-cpython-313/torch/distributed/launcher copying torch/distributed/launcher/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/launcher copying torch/distributed/launcher/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/launcher creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms copying torch/distributed/algorithms/join.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms copying torch/distributed/algorithms/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms creating build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/random.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_tp_conv.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_collective_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_dispatch.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/device_mesh.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_redistribute.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/placement_types.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_sharding_prop.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor copying torch/distributed/_tensor/_op_schema.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor creating build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/batch_dim_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/config.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/experimental_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/graph_optimization.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/iter_graph_module.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/data_parallel.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/log_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/graph_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/partial_lower.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/parallel_mode.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/gm_transformation.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/distribute.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd copying torch/distributed/_spmd/comm_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_spmd creating build/lib.linux-x86_64-cpython-313/torch/distributed/_composable copying torch/distributed/_composable/checkpoint_activation.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable copying torch/distributed/_composable/fully_shard.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable copying torch/distributed/_composable/replicate.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable copying torch/distributed/_composable/contract.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable copying torch/distributed/_composable/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable creating build/lib.linux-x86_64-cpython-313/torch/distributed/tensor copying torch/distributed/tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic copying torch/distributed/elastic/control_plane.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic copying torch/distributed/elastic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic creating build/lib.linux-x86_64-cpython-313/torch/distributed/_tools copying torch/distributed/_tools/memory_tracker.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tools copying torch/distributed/_tools/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tools creating build/lib.linux-x86_64-cpython-313/torch/distributed/nn copying torch/distributed/nn/functional.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn copying torch/distributed/nn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn creating build/lib.linux-x86_64-cpython-313/torch/distributed/_sharding_spec copying torch/distributed/_sharding_spec/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_sharding_spec creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_plan copying torch/distributed/_shard/sharding_plan/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_plan copying torch/distributed/_shard/sharding_plan/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_plan creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_optim copying torch/distributed/_shard/sharded_optim/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_optim copying torch/distributed/_shard/sharded_optim/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_optim creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/shard.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/logger.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/metadata.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/reshard.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/logging_handlers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor copying torch/distributed/_shard/sharded_tensor/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/checkpoint copying torch/distributed/_shard/checkpoint/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/checkpoint creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec copying torch/distributed/_shard/sharding_spec/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec copying torch/distributed/_shard/sharding_spec/chunk_sharding_spec.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec copying torch/distributed/_shard/sharding_spec/_internals.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec copying torch/distributed/_shard/sharding_spec/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops copying torch/distributed/_shard/sharded_tensor/_ops/misc_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops copying torch/distributed/_shard/sharded_tensor/_ops/binary_cmp.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops copying torch/distributed/_shard/sharded_tensor/_ops/_common.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops copying torch/distributed/_shard/sharded_tensor/_ops/tensor_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops copying torch/distributed/_shard/sharded_tensor/_ops/init.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops copying torch/distributed/_shard/sharded_tensor/_ops/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharded_tensor/_ops creating build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops copying torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops/_common.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops copying torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops/embedding.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops copying torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops copying torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops/embedding_bag.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops creating build/lib.linux-x86_64-cpython-313/torch/distributed/rpc/_testing copying torch/distributed/rpc/_testing/faulty_agent_backend_registry.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc/_testing copying torch/distributed/rpc/_testing/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/rpc/_testing creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/default_hooks.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/mixed_precision_hooks.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/post_localSGD_hook.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/ddp_zero_hook.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/optimizer_overlap_hooks.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/powerSGD_hook.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/debugging_hooks.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/quantization_hooks.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks copying torch/distributed/algorithms/ddp_comm_hooks/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/ddp_comm_hooks creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_quantization copying torch/distributed/algorithms/_quantization/quantization.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_quantization copying torch/distributed/algorithms/_quantization/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_quantization creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_comm_hooks copying torch/distributed/algorithms/_comm_hooks/default_hooks.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_comm_hooks copying torch/distributed/algorithms/_comm_hooks/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_comm_hooks creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_optimizer_overlap copying torch/distributed/algorithms/_optimizer_overlap/optimizer_overlap.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_optimizer_overlap copying torch/distributed/algorithms/_optimizer_overlap/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_optimizer_overlap creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/model_averaging copying torch/distributed/algorithms/model_averaging/averagers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/model_averaging copying torch/distributed/algorithms/model_averaging/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/model_averaging copying torch/distributed/algorithms/model_averaging/hierarchical_model_averager.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/model_averaging copying torch/distributed/algorithms/model_averaging/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/model_averaging creating build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_checkpoint copying torch/distributed/algorithms/_checkpoint/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_checkpoint copying torch/distributed/algorithms/_checkpoint/checkpoint_wrapper.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/algorithms/_checkpoint creating build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/random_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/pointwise_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/conv_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/experimental_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/math_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/basic_strategy.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/view_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/embedding_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/common_rules.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/matrix_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/tensor_ops.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops copying torch/distributed/_tensor/ops/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/ops creating build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/experimental copying torch/distributed/_tensor/experimental/local_map.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/experimental copying torch/distributed/_tensor/experimental/tp_transform.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/experimental copying torch/distributed/_tensor/experimental/attention.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/experimental copying torch/distributed/_tensor/experimental/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/experimental creating build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/debug copying torch/distributed/_tensor/debug/visualize_sharding.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/debug copying torch/distributed/_tensor/debug/_op_coverage.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/debug copying torch/distributed/_tensor/debug/comm_mode.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/debug copying torch/distributed/_tensor/debug/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_tensor/debug creating build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_init.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_common.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_collectives.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/fully_shard.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_param.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_param_group.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/_fsdp_state.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp copying torch/distributed/_composable/fsdp/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/_composable/fsdp creating build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/loss.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/_data_parallel_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/input_reshard.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/fsdp.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/ddp.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/style.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel copying torch/distributed/tensor/parallel/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/tensor/parallel creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/static_tcp_rendezvous.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/etcd_rendezvous.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/utils.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/etcd_store.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/etcd_rendezvous_backend.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/registry.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/dynamic_rendezvous.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/c10d_rendezvous_backend.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous copying torch/distributed/elastic/rendezvous/etcd_server.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/rendezvous creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent copying torch/distributed/elastic/agent/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/timer copying torch/distributed/elastic/timer/file_based_local_timer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/timer copying torch/distributed/elastic/timer/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/timer copying torch/distributed/elastic/timer/debug_info_logging.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/timer copying torch/distributed/elastic/timer/local_timer.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/timer copying torch/distributed/elastic/timer/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/timer creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/metrics copying torch/distributed/elastic/metrics/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/metrics copying torch/distributed/elastic/metrics/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/metrics creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils copying torch/distributed/elastic/utils/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils copying torch/distributed/elastic/utils/logging.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils copying torch/distributed/elastic/utils/log_level.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils copying torch/distributed/elastic/utils/store.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils copying torch/distributed/elastic/utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils copying torch/distributed/elastic/utils/distributed.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/events copying torch/distributed/elastic/events/handlers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/events copying torch/distributed/elastic/events/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/events copying torch/distributed/elastic/events/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/events creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing copying torch/distributed/elastic/multiprocessing/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing copying torch/distributed/elastic/multiprocessing/tail_log.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing copying torch/distributed/elastic/multiprocessing/redirects.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing copying torch/distributed/elastic/multiprocessing/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent/server copying torch/distributed/elastic/agent/server/api.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent/server copying torch/distributed/elastic/agent/server/health_check_server.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent/server copying torch/distributed/elastic/agent/server/local_elastic_agent.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent/server copying torch/distributed/elastic/agent/server/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/agent/server creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils/data copying torch/distributed/elastic/utils/data/elastic_distributed_sampler.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils/data copying torch/distributed/elastic/utils/data/cycling_iterator.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils/data copying torch/distributed/elastic/utils/data/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/utils/data creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/errors copying torch/distributed/elastic/multiprocessing/errors/handlers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/errors copying torch/distributed/elastic/multiprocessing/errors/error_handler.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/errors copying torch/distributed/elastic/multiprocessing/errors/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/errors creating build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/subprocess_handler copying torch/distributed/elastic/multiprocessing/subprocess_handler/handlers.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/subprocess_handler copying torch/distributed/elastic/multiprocessing/subprocess_handler/subprocess_handler.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/subprocess_handler copying torch/distributed/elastic/multiprocessing/subprocess_handler/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/elastic/multiprocessing/subprocess_handler creating build/lib.linux-x86_64-cpython-313/torch/distributed/nn/jit copying torch/distributed/nn/jit/instantiator.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn/jit copying torch/distributed/nn/jit/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn/jit creating build/lib.linux-x86_64-cpython-313/torch/distributed/nn/api copying torch/distributed/nn/api/remote_module.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn/api copying torch/distributed/nn/api/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn/api creating build/lib.linux-x86_64-cpython-313/torch/distributed/nn/jit/templates copying torch/distributed/nn/jit/templates/remote_module_template.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn/jit/templates copying torch/distributed/nn/jit/templates/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/distributed/nn/jit/templates creating build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/accelerator_partitioner.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/_sym_dispatch_mode.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/refinement_types.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/_config.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/partitioner_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/validator.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/const_fold.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/schema_type_annotation.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/rewriter.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/merge_matmul.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/sym_node.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/graph_gradual_typechecker.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/recording.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/unify_refinements.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/symbolic_shapes.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/optimization.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/debug.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/proxy_tensor.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/_backward_state.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/meta_tracer.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental copying torch/fx/experimental/normalize.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental creating build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/annotate_getitem_nodes.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/shape_prop.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/pass_manager.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/operator_support.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/split_module.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/net_min_base.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/reinplace.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/split_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/graph_drawer.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/param_fetch.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/graph_manipulation.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/graph_transform_observer.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/splitter_base.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/fake_tensor_prop.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/tools_common.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes copying torch/fx/passes/runtime_assert.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes creating build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/dispatch.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/core.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/unification_tools.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/match.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/variable.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification copying torch/fx/experimental/unification/more.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification creating build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/constraint.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/constraint_generator.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/constraint_transformation.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/operation.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/transform_to_z3.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/z3_types.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/util.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types copying torch/fx/experimental/migrate_gradual_types/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/migrate_gradual_types creating build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch copying torch/fx/experimental/unification/multipledispatch/variadic.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch copying torch/fx/experimental/unification/multipledispatch/core.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch copying torch/fx/experimental/unification/multipledispatch/utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch copying torch/fx/experimental/unification/multipledispatch/conflict.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch copying torch/fx/experimental/unification/multipledispatch/dispatcher.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch copying torch/fx/experimental/unification/multipledispatch/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/experimental/unification/multipledispatch creating build/lib.linux-x86_64-cpython-313/torch/fx/passes/infra copying torch/fx/passes/infra/pass_manager.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/infra copying torch/fx/passes/infra/partitioner.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/infra copying torch/fx/passes/infra/pass_base.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/infra copying torch/fx/passes/infra/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/infra creating build/lib.linux-x86_64-cpython-313/torch/fx/passes/backends copying torch/fx/passes/backends/cudagraphs.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/backends copying torch/fx/passes/backends/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/backends creating build/lib.linux-x86_64-cpython-313/torch/fx/passes/tests copying torch/fx/passes/tests/test_pass_manager.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/tests copying torch/fx/passes/tests/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/tests creating build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils copying torch/fx/passes/utils/common.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils copying torch/fx/passes/utils/source_matcher_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils copying torch/fx/passes/utils/matcher_with_name_node_map_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils copying torch/fx/passes/utils/fuser_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils copying torch/fx/passes/utils/matcher_utils.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils copying torch/fx/passes/utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/utils creating build/lib.linux-x86_64-cpython-313/torch/fx/passes/dialect copying torch/fx/passes/dialect/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/dialect creating build/lib.linux-x86_64-cpython-313/torch/fx/passes/dialect/common copying torch/fx/passes/dialect/common/cse_pass.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/dialect/common copying torch/fx/passes/dialect/common/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/fx/passes/dialect/common creating build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/functional_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/input_output_analysis.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/traced_function_transforms.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/subclass_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/logging_utils.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/jit_compile_runtime_wrappers.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/autograd_cache.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/runtime_wrappers.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/dispatch_and_compile_graph.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/collect_metadata_analysis.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/schemas.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd copying torch/_functorch/_aot_autograd/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_functorch/_aot_autograd creating build/lib.linux-x86_64-cpython-313/torch/_refs/special copying torch/_refs/special/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_refs/special creating build/lib.linux-x86_64-cpython-313/torch/_refs/linalg copying torch/_refs/linalg/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_refs/linalg creating build/lib.linux-x86_64-cpython-313/torch/_refs/nn copying torch/_refs/nn/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_refs/nn creating build/lib.linux-x86_64-cpython-313/torch/_refs/nn/functional copying torch/_refs/nn/functional/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_refs/nn/functional creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/onnx_proto_utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/jit_utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/onnxruntime.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/exporter.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/io_adapter.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/registration.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/_beartype.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal copying torch/onnx/_internal/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics copying torch/onnx/_internal/diagnostics/_diagnostic.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics copying torch/onnx/_internal/diagnostics/_rules.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics copying torch/onnx/_internal/diagnostics/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/op_validation.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/decomposition_skip.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/dynamo_graph_extractor.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/serialization.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/fx_onnx_interpreter.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/diagnostics.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/torch_export_graph_extractor.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/registration.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/onnxfunction_dispatcher.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/type_utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/patcher.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/_pass.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/fx_symbolic_graph_extractor.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx copying torch/onnx/_internal/fx/decomposition_table.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra copying torch/onnx/_internal/diagnostics/infra/utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra copying torch/onnx/_internal/diagnostics/infra/context.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra copying torch/onnx/_internal/diagnostics/infra/_infra.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra copying torch/onnx/_internal/diagnostics/infra/decorator.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra copying torch/onnx/_internal/diagnostics/infra/formatter.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra copying torch/onnx/_internal/diagnostics/infra/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_reporting_configuration.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_external_property_file_reference.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_thread_flow_location.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_region.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_node.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_artifact_location.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_conversion.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_special_locations.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_invocation.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_fix.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_external_properties.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_edge.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_location_relationship.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_web_request.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_multiformat_message_string.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_result.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_property_bag.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_tool_component_reference.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_sarif_log.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_thread_flow.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/version.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_run_automation_details.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_message.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_graph.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_version_control_details.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_artifact_change.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_graph_traversal.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_reporting_descriptor_reference.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_exception.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_reporting_descriptor_relationship.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_result_provenance.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_translation_metadata.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_web_response.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_suppression.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_stack_frame.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_address.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_configuration_override.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_external_property_file_references.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_reporting_descriptor.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_logical_location.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_stack.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_edge_traversal.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_artifact_content.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_notification.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_artifact.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_code_flow.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_replacement.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_run.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_tool_component.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_location.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_rectangle.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_tool.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_physical_location.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif copying torch/onnx/_internal/diagnostics/infra/sarif/_attachment.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/diagnostics/infra/sarif creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/readability.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/functionalization.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/decomp.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/modularization.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/type_promotion.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/virtualization.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes copying torch/onnx/_internal/fx/passes/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/passes creating build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/analysis copying torch/onnx/_internal/fx/analysis/unsupported_nodes.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/analysis copying torch/onnx/_internal/fx/analysis/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/onnx/_internal/fx/analysis creating build/lib.linux-x86_64-cpython-313/torch/signal/windows copying torch/signal/windows/windows.py -> build/lib.linux-x86_64-cpython-313/torch/signal/windows copying torch/signal/windows/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/signal/windows creating build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/scatter_gather.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/data_parallel.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/_functions.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/parallel_apply.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/replicate.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/comm.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel copying torch/nn/parallel/distributed.py -> build/lib.linux-x86_64-cpython-313/torch/nn/parallel creating build/lib.linux-x86_64-cpython-313/torch/nn/attention copying torch/nn/attention/_flex_attention.py -> build/lib.linux-x86_64-cpython-313/torch/nn/attention copying torch/nn/attention/bias.py -> build/lib.linux-x86_64-cpython-313/torch/nn/attention copying torch/nn/attention/_utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/attention copying torch/nn/attention/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/attention creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic copying torch/nn/intrinsic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic creating build/lib.linux-x86_64-cpython-313/torch/nn/backends copying torch/nn/backends/thnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/backends copying torch/nn/backends/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/backends creating build/lib.linux-x86_64-cpython-313/torch/nn/quantized copying torch/nn/quantized/functional.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized copying torch/nn/quantized/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized creating build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/_deprecation_utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/fusion.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/weight_norm.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/parametrizations.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/convert_parameters.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/clip_grad.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/stateless.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/_per_sample_grad.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/parametrize.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/_named_member_accessor.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/prune.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/init.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/spectral_norm.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils copying torch/nn/utils/memory_format.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils creating build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/dropout.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/distance.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/upsampling.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/loss.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/instancenorm.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/channelshuffle.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/normalization.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/lazy.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/activation.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/batchnorm.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/module.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/_functions.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/sparse.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/adaptive.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/transformer.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/fold.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/pooling.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/padding.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/container.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/flatten.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/pixelshuffle.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules copying torch/nn/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/qat copying torch/nn/qat/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat creating build/lib.linux-x86_64-cpython-313/torch/nn/quantizable copying torch/nn/quantizable/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantizable creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized copying torch/nn/intrinsic/quantized/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/modules copying torch/nn/intrinsic/modules/fused.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/modules copying torch/nn/intrinsic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat copying torch/nn/intrinsic/qat/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/modules copying torch/nn/intrinsic/quantized/modules/conv_relu.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/modules copying torch/nn/intrinsic/quantized/modules/bn_relu.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/modules copying torch/nn/intrinsic/quantized/modules/linear_relu.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/modules copying torch/nn/intrinsic/quantized/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/dynamic copying torch/nn/intrinsic/quantized/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/dynamic creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/dynamic/modules copying torch/nn/intrinsic/quantized/dynamic/modules/linear_relu.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/dynamic/modules copying torch/nn/intrinsic/quantized/dynamic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/quantized/dynamic/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat/modules copying torch/nn/intrinsic/qat/modules/linear_fused.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat/modules copying torch/nn/intrinsic/qat/modules/linear_relu.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat/modules copying torch/nn/intrinsic/qat/modules/conv_fused.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat/modules copying torch/nn/intrinsic/qat/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/intrinsic/qat/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference copying torch/nn/quantized/_reference/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference creating build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/dropout.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/normalization.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/embedding_ops.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/activation.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/batchnorm.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/functional_modules.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules copying torch/nn/quantized/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic copying torch/nn/quantized/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic creating build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules copying torch/nn/quantized/_reference/modules/utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules copying torch/nn/quantized/_reference/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules copying torch/nn/quantized/_reference/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules copying torch/nn/quantized/_reference/modules/sparse.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules copying torch/nn/quantized/_reference/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules copying torch/nn/quantized/_reference/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/_reference/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic/modules copying torch/nn/quantized/dynamic/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic/modules copying torch/nn/quantized/dynamic/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic/modules copying torch/nn/quantized/dynamic/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic/modules copying torch/nn/quantized/dynamic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantized/dynamic/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/conv_expanded_weights.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/expanded_weights_impl.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/group_norm_expanded_weights.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/embedding_expanded_weights.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/instance_norm_expanded_weights.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/conv_utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/layer_norm_expanded_weights.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/linear_expanded_weights.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights copying torch/nn/utils/_expanded_weights/expanded_weights_utils.py -> build/lib.linux-x86_64-cpython-313/torch/nn/utils/_expanded_weights creating build/lib.linux-x86_64-cpython-313/torch/nn/qat/modules copying torch/nn/qat/modules/conv.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/modules copying torch/nn/qat/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/modules copying torch/nn/qat/modules/embedding_ops.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/modules copying torch/nn/qat/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/qat/dynamic copying torch/nn/qat/dynamic/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/dynamic creating build/lib.linux-x86_64-cpython-313/torch/nn/qat/dynamic/modules copying torch/nn/qat/dynamic/modules/linear.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/dynamic/modules copying torch/nn/qat/dynamic/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/qat/dynamic/modules creating build/lib.linux-x86_64-cpython-313/torch/nn/quantizable/modules copying torch/nn/quantizable/modules/activation.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantizable/modules copying torch/nn/quantizable/modules/rnn.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantizable/modules copying torch/nn/quantizable/modules/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/nn/quantizable/modules creating build/lib.linux-x86_64-cpython-313/torch/_export/db copying torch/_export/db/logging.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db copying torch/_export/db/case.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db copying torch/_export/db/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db copying torch/_export/db/gen_example.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db creating build/lib.linux-x86_64-cpython-313/torch/_export/pass_infra copying torch/_export/pass_infra/proxy_value.py -> build/lib.linux-x86_64-cpython-313/torch/_export/pass_infra copying torch/_export/pass_infra/node_metadata.py -> build/lib.linux-x86_64-cpython-313/torch/_export/pass_infra copying torch/_export/pass_infra/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_export/pass_infra creating build/lib.linux-x86_64-cpython-313/torch/_export/serde copying torch/_export/serde/schema.py -> build/lib.linux-x86_64-cpython-313/torch/_export/serde copying torch/_export/serde/serialize.py -> build/lib.linux-x86_64-cpython-313/torch/_export/serde copying torch/_export/serde/union.py -> build/lib.linux-x86_64-cpython-313/torch/_export/serde copying torch/_export/serde/schema_check.py -> build/lib.linux-x86_64-cpython-313/torch/_export/serde copying torch/_export/serde/upgrade.py -> build/lib.linux-x86_64-cpython-313/torch/_export/serde copying torch/_export/serde/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_export/serde creating build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/remove_runtime_assertions.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/lift_constants_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/replace_set_grad_with_hop_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/_node_metadata_hook.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/replace_sym_size_ops_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/add_runtime_assertions_for_constraints_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/functionalize_side_effectful_ops_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/collect_tracepoints_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/replace_view_ops_with_view_copy_ops_pass.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/constant_folding.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes copying torch/_export/passes/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_export/passes creating build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/fn_with_kwargs.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/cond_operands.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/cond_closed_over_variable.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_if_guard.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/optional_input.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/cond_branch_nonlocal_variables.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/user_input_mutation.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_constructor.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/list_contains.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/type_reflection_method.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_assert.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/cond_predicate.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_map.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/cond_branch_nested_function.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/specialized_attribute.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/null_context_manager.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/tensor_setattr.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_round.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/nested_function.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/cond_branch_class_method.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/class_method.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/static_for_loop.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/static_if.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/constrain_as_value_example.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_slicing.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/torch_sym_min.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/pytree_flatten.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dynamic_shape_view.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/scalar_output.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/decorator.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/dictionary.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/model_attr_mutation.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/assume_constant_result.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/list_unpack.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/autograd_function.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/__init__.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples copying torch/_export/db/examples/constrain_as_size_example.py -> build/lib.linux-x86_64-cpython-313/torch/_export/db/examples creating build/lib.linux-x86_64-cpython-313/torchgen/packaged creating build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen creating build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/native copying torchgen/packaged/ATen/native/native_functions.yaml -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/native copying torchgen/packaged/ATen/native/tags.yaml -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/native creating build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/DispatchKeyFunction.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/UnboxingFunctions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/ATenOpList.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/NativeMetaFunctions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RedispatchFunctions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/aten_interned_strings.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/DispatchKeyNativeFunctions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/UfuncCPUKernel.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/NativeFunctions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/NativeMetaFunction.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/Operators.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/TensorBody.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/UnboxingFunctions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/TensorMethods.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/NativeFunction.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/Operators.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/Operator.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegisterBackendSelect.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/Function.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/CompositeViewCopyKernels.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RedispatchFunctions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegisterFunctionalization.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/Functions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/LazyNonNativeIr.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/DispatchKeyFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/UfuncCUDA.cu -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegisterDispatchKey.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegisterSchema.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/Functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegisterDispatchDefinitions.ini -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/DispatchKeyNativeFunctions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/LazyIr.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegistrationDeclarations.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/DispatchKeyFunctions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/UfuncCPU.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/enum_tag.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/RegisterCodegenUnboxedKernels.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/FunctionalInverses.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates copying torchgen/packaged/ATen/templates/MethodOperators.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/ATen/templates creating build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_variable_type.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_trace_type.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_autograd.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_view_funcs.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_inplace_or_view_type.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/build.bzl -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_annotated_fn_args.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/BUILD.bazel -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/load_derivatives.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_autograd_functions.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_python_functions.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/context.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/derivatives.yaml -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/gen_variable_factories.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/README.md -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/deprecated.yaml -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd copying torchgen/packaged/autograd/__init__.py -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd creating build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/VariableType.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/ADInplaceOrViewType.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_linalg_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_return_types.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_sparse_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_fft_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/ViewFuncs.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/VariableType.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_variable_methods.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_torch_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/ViewFuncs.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_nested_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_enum_tag.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/Functions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_special_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_return_types.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/TraceType.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/variable_factories.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/annotated_fn_args.py.in -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/Functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_functions.h -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torchgen/packaged/autograd/templates/python_nn_functions.cpp -> build/lib.linux-x86_64-cpython-313/torchgen/packaged/autograd/templates copying torch/_VF.pyi -> build/lib.linux-x86_64-cpython-313/torch copying torch/return_types.pyi -> build/lib.linux-x86_64-cpython-313/torch copying torch/py.typed -> build/lib.linux-x86_64-cpython-313/torch creating build/lib.linux-x86_64-cpython-313/torch/bin copying torch/bin/torch_shm_manager -> build/lib.linux-x86_64-cpython-313/torch/bin creating build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_distributed_autograd.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_onnx.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_distributed_rpc.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_lazy_ts_backend.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_distributed_rpc_testing.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_nvtx.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_verbose.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_aoti.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_cudnn.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_itt.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_autograd.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_functions.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_monitor.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_lazy.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_functorch.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_profiler.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_distributed_c10d.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_cpu.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/__init__.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_VariableFunctions.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/_C/_nn.pyi -> build/lib.linux-x86_64-cpython-313/torch/_C copying torch/fx/__init__.pyi -> build/lib.linux-x86_64-cpython-313/torch/fx copying torch/nn/parameter.pyi -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/nn/functional.pyi -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/utils/data/datapipes/datapipe.pyi -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes creating build/lib.linux-x86_64-cpython-313/torch/include copying torch/include/libshm.h -> build/lib.linux-x86_64-cpython-313/torch/include creating build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ATen.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/AccumulateType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ArrayRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Backend.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Backtrace.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/BlasBackend.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CPUApplyUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CPUFixedAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CPUGeneratorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CachedTensorUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CollapseDims.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Config.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Context.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/DLConvertor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Device.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/DeviceAccelerator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/DeviceGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/DimVector.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Dimname.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Dispatch_v2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/DynamicLibrary.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/EmptyTensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ExpandBase.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ExpandUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Formatting.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/FuncTorchTLS.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/FunctionalStorageImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/FunctionalTensorWrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Generator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/InferSize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/InitialTensorOptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Layout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/LegacyBatchedFallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/LegacyBatchedTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/LegacyVmapMode.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/LegacyVmapTransforms.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/LinalgBackend.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/MapAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/MatrixRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/MemoryOverlap.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/NamedTensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/NamedTensorUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/NestedTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/NumericUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/OpMathType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/OpaqueTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/PTThreadPool.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/PadNd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Parallel-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Parallel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ParallelFuture.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ParallelNative.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ParallelOpenMP.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/PythonTorchFunctionTLS.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/SavedTensorHooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Scalar.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ScalarOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ScalarType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/SequenceNumber.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/SmallVector.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/SparseCsrTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/SparseCsrTensorUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/SparseTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Storage.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/StorageUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorAccessor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorGeometry.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorIndexing.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorIterator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorIteratorInternal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorMeta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorNames.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorOperators.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorOptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorSubclassLikeUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TensorUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ThreadLocalPythonObjects.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ThreadLocalState.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TracerMode.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/TypeDefault.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Version.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/WrapDimUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/WrapDimUtilsMulti.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/autocast_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/ceil_div.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/code_template.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/cpp_custom_type_hack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/div_rtn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/dlpack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/jit_macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/jiterator_macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/record_function.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CPUFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CPUFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeExplicitAutogradFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeExplicitAutogradFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeExplicitAutogradNonFunctionalFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeExplicitAutogradNonFunctionalFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeImplicitAutogradFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeImplicitAutogradFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeImplicitAutogradNestedTensorFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CompositeImplicitAutogradNestedTensorFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/MetaFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/MetaFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/MethodOperators.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/NativeFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/NativeMetaFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/Operators.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/RedispatchFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/RegistrationDeclarations.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/VmapGeneratedPlumbing.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CUDAFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen copying torch/include/ATen/CUDAFunctions_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu copying torch/include/ATen/cpu/FlushDenormal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu copying torch/include/ATen/cpu/Utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu copying torch/include/ATen/cpu/vml.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/missing_vld1_neon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/missing_vst1_neon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_bfloat16.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_complex_double.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_complex_float.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_convert.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_double.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_float.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_float_neon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_half_neon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_int.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_mask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 copying torch/include/ATen/cpu/vec/vec256/vec256_qint.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256 creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_bfloat16_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_common_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_complex_double_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_complex_float_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_double_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_float_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_int16_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_int32_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_int64_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_qint32_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_qint8_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vec256_quint8_vsx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx copying torch/include/ATen/cpu/vec/vec256/vsx/vsx_helpers.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/vsx creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/zarch copying torch/include/ATen/cpu/vec/vec256/zarch/vec256_zarch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec256/zarch creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_bfloat16.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_complex_double.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_complex_float.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_convert.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_double.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_float.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_int.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_mask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/vec512/vec512_qint.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec/vec512 copying torch/include/ATen/cpu/vec/functional.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/functional_base.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/functional_bfloat16.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/intrinsics.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/vec.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/vec_base.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/vec_convert.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/vec_half.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/vec_mask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec copying torch/include/ATen/cpu/vec/vec_n.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cpu/vec creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ATenGeneral.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ATenOpList.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ATen_fwd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ATen_pch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Array.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Backtrace.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/CachingHostAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/CheckMemoryFormat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/DeprecatedTypeProperties.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/DeprecatedTypePropertiesRegistry.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Dict.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Dict_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/DimVector.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Dimname.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/DistributionsHelper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Formatting.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Generator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/GeneratorForPrivateuseone.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/IListRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/IListRef_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/LegacyTypeDispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/List.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/List_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/MT19937RNGEngine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/NamedTensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/NestedIntSymNodeImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/PhiloxRNGEngine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/PythonFallbackKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/PythonOpRegistrationTrampoline.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/QuantizerBase.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Range.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Reduction.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Scalar.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ScalarType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/TensorAccessor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/TensorBase.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/TorchDispatchUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/TransformationHelper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/UndefinedTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/UnsafeFromTH.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/VariableHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Variadic.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/Vitals.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/alias_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/blob.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/builtin_function.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/class_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/custom_class.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/dynamic_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/enum_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/function.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/function_schema.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/function_schema_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/functional.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/grad_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/interned_strings.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/interned_strings_class.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ivalue.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ivalue_inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/ivalue_to.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/jit_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/jit_type_base.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/operator_name.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/qualified_name.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/rref_interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/stack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/symbol.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/type_factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/type_ptr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/typeid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/TensorBody.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/aten_interned_strings.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core copying torch/include/ATen/core/enum_tag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/ApplyGridUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/AsmUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/Atomic.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAApplyUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAGraphsUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDATensorMethods.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/DeviceUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/NumericLimits.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/PhiloxUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/ScanUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/cub.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/cub_definitions.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/ATenCUDAGeneral.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDABlas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAConfig.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAContext.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAContextLight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDADataType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDADevice.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAEvent.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAGeneratorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAGraph.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDASparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDASparseBlas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDASparseDescriptors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CUDAUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/CachingHostAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/EmptyTensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/Exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/PeerToPeerAccess.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/PhiloxCudaState.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/PinnedMemoryAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/Sleep.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/ThrustAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/cub.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/jiterator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/jiterator_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda copying torch/include/ATen/cuda/llvm_jit_strings.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/IndexUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/IntegerDivider.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/OffsetCalculator.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/PhiloxCudaStateRaw.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/TensorInfo.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/UnpackRaw.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/CUDAHooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/DeviceThreadHandles.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/KernelUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail copying torch/include/ATen/cuda/detail/LazyNVRTC.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/detail creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/GemmCommon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/GemmHipblaslt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/GemmRocblas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/StreamTimer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/Tunable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/TunableGemm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable copying torch/include/ATen/cuda/tunable/TunableOp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cuda/tunable creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/Descriptors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/Exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/Handle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/Handles.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/Types.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/Utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn copying torch/include/ATen/cudnn/cudnn-wrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/cudnn creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/ADInterpreters.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/BatchRulesHelper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/BatchedFallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/BatchedTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/BatchingMetaprogramming.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/DynamicLayer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/FunctionalizeInterpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/Interpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/LegacyVmapTransforms.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/Macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/PlumbingHelper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/TensorWrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch copying torch/include/ATen/functorch/VmapInterpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/functorch creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/from_blob.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_adaptive_avg_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_batch_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_batch_dim_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_batch_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_batch_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_relu_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_relu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_relu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_add_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_addmm_activation_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_aminmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_aminmax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_aminmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_aminmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_aminmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_aminmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_foreach_non_finite_check_and_unscale.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_foreach_non_finite_check_and_unscale_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_foreach_non_finite_check_and_unscale_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_foreach_non_finite_check_and_unscale_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_foreach_non_finite_check_and_unscale_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_foreach_non_finite_check_and_unscale_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_amp_update_scale_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_async.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_async_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_async_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_async_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_async_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_scalar.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_scalar_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_scalar_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_scalar_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_tensor_metadata.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_tensor_metadata_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_tensor_metadata_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_assert_tensor_metadata_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_full_precision.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_full_precision_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_full_precision_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_full_precision_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_reduced_precision.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_reduced_precision_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_reduced_precision_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_autocast_to_reduced_precision_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_impl_index_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_no_update.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_no_update_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_no_update_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_no_update_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_with_update.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_with_update_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_with_update_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_with_update_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_with_update_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_batch_norm_with_update_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Byte.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Byte_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Byte_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Byte_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Char.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Char_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Char_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Char_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Double.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Double_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Double_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Double_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Float.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Float_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Float_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Float_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Half.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Half_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Half_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Half_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Int.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Int_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Int_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Int_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Long.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Long_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Long_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Long_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Short.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Short_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Short_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cast_Short_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_forward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cdist_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cholesky_solve_helper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cholesky_solve_helper_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cholesky_solve_helper_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cholesky_solve_helper_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cholesky_solve_helper_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cholesky_solve_helper_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_choose_qparams_per_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_choose_qparams_per_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_choose_qparams_per_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_choose_qparams_per_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_chunk_cat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_chunk_cat_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_chunk_cat_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_chunk_cat_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_chunk_cat_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesce_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesce_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesce_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesced.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesced_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesced_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesced_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_coalesced_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_compute_linear_combination.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_compute_linear_combination_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_compute_linear_combination_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_compute_linear_combination_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_compute_linear_combination_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_physical.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_physical_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_physical_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conj_physical_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conv_depthwise2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conv_depthwise2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conv_depthwise2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_conv_depthwise2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_coo_to_csr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_indices_from_csr_to_coo_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_weight_to_int4pack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_weight_to_int4pack_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_weight_to_int4pack_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_weight_to_int4pack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convert_weight_to_int4pack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_double_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_double_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_double_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_double_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_mode_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_mode_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_mode_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_and_resize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_and_resize_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_and_resize_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_and_resize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_copy_from_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_compress.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_compress_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_compress_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_compress_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_search.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_search_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_search_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cslt_sparse_mm_search_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_ctc_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_ctc_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_ctc_loss_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_ctc_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_ctc_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_ctc_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_init_dropout_state.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_init_dropout_state_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_init_dropout_state_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_init_dropout_state_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_init_dropout_state_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_flatten_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_flatten_weight_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_flatten_weight_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_flatten_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_flatten_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cudnn_rnn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_clear_plan_cache.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_clear_plan_cache_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_clear_plan_cache_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_clear_plan_cache_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_max_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_max_size_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_max_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_max_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_size_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_get_plan_cache_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_set_plan_cache_max_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_set_plan_cache_max_size_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_set_plan_cache_max_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cufft_set_plan_cache_max_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummax_helper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummax_helper_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummax_helper_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummax_helper_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummax_helper_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummin_helper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummin_helper_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummin_helper_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummin_helper_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_cummin_helper_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_debug_has_internal_overlap.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_debug_has_internal_overlap_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_debug_has_internal_overlap_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_debug_has_internal_overlap_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dimI.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dimI_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dimI_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dimV.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dimV_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dimV_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dim_arange.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dim_arange_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dim_arange_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dim_arange_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dirichlet_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dirichlet_grad_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dirichlet_grad_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dirichlet_grad_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dirichlet_grad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_dirichlet_grad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficient_attention_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_efficientzerotensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_dense_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_dense_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_dense_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_dense_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_dense_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_dense_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_forward_only.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_forward_only_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_forward_only_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_forward_only_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_forward_only_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_forward_only_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_per_sample_weights_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_per_sample_weights_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_per_sample_weights_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_per_sample_weights_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_per_sample_weights_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_per_sample_weights_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_sparse_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_sparse_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_sparse_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_embedding_bag_sparse_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_affine_quantized.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_affine_quantized_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_affine_quantized_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_affine_quantized_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_affine_quantized_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_per_channel_affine_quantized.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_per_channel_affine_quantized_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_per_channel_affine_quantized_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_per_channel_affine_quantized_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_empty_per_channel_affine_quantized_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_euclidean_dist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_euclidean_dist_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_euclidean_dist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_euclidean_dist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_channel_affine_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_learnable_per_tensor_affine_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2c.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2c_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2c_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2c_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2c_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2r.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2r_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2r_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2r_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_c2r_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_r2c.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_r2c_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_r2c_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_r2c_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fft_r2c_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fill_mem_eff_dropout_mask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fill_mem_eff_dropout_mask_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fill_mem_eff_dropout_mask_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fill_mem_eff_dropout_mask_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fill_mem_eff_dropout_mask_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_flash_attention_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foobar.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foobar_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foobar_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foobar_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foobar_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_abs.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_abs_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_abs_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_abs_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_abs_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_acos.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_acos_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_acos_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_acos_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_acos_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_add.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_add_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_add_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_add_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_add_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcdiv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcdiv_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcdiv_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcdiv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcdiv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcmul_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcmul_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_addcmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_asin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_asin_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_asin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_asin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_asin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_atan.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_atan_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_atan_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_atan_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_atan_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_ceil.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_ceil_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_ceil_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_ceil_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_ceil_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_max.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_max_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_max_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_max_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_max_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_min.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_min_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_min_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_min_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_clamp_min_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_copy_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cos.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cos_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cos_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cos_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cos_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cosh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cosh_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cosh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cosh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_cosh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_div.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_div_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_div_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_div_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_div_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erf_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erf_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erfc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erfc_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erfc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erfc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_erfc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_exp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_exp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_exp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_exp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_exp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_expm1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_expm1_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_expm1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_expm1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_expm1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_floor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_floor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_floor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_floor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_floor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_frac.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_frac_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_frac_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_frac_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_frac_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lerp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lerp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lerp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lerp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lerp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lgamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lgamma_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lgamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lgamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_lgamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log10.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log10_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log10_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log10_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log10_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log1p.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log1p_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log1p_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log1p_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log1p_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log2_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log2_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_log_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_max.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_max_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_max_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_max_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_max_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_maximum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_maximum_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_maximum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_maximum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_maximum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_minimum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_minimum_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_minimum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_minimum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_minimum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_mul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_mul_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_mul_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_mul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_mul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_neg.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_neg_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_neg_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_neg_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_neg_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_pow.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_pow_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_pow_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_pow_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_pow_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_reciprocal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_reciprocal_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_reciprocal_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_reciprocal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_reciprocal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_round.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_round_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_round_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_round_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_round_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sigmoid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sigmoid_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sigmoid_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sigmoid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sigmoid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sign.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sign_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sign_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sign_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sign_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sin_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sinh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sinh_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sinh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sinh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sinh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sqrt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sqrt_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sqrt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sqrt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sqrt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sub.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sub_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sub_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sub_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_sub_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tan.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tan_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tan_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tan_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tan_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tanh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tanh_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tanh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tanh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_tanh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_trunc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_trunc_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_trunc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_trunc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_trunc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_zero.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_zero_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_zero_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_zero_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_foreach_zero_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_async.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_async_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_async_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_async_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_scalar.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_scalar_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_scalar_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_assert_scalar_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_for_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_for_size_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_for_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_for_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_functional_sym_constrain_range_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adagrad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adagrad_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adagrad_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adagrad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adagrad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adam.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adam_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adam_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adam_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adam_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adam_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adamw.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adamw_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adamw_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adamw_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adamw_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_adamw_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_dropout_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_dropout_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_dropout_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_dropout_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_moving_avg_obs_fq_helper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_moving_avg_obs_fq_helper_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_moving_avg_obs_fq_helper_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_moving_avg_obs_fq_helper_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_moving_avg_obs_fq_helper_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_moving_avg_obs_fq_helper_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sdp_choice.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sdp_choice_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sdp_choice_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sdp_choice_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sdp_choice_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sdp_choice_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sgd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sgd_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sgd_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sgd_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sgd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fused_sgd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_fw_primal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_gather_sparse_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_gather_sparse_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_gather_sparse_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_gather_sparse_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_grid_sampler_2d_cpu_fallback_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_compatible_shallow_copy_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_compatible_shallow_copy_type_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_compatible_shallow_copy_type_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_compatible_shallow_copy_type_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_same_storage_numel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_same_storage_numel_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_same_storage_numel_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_has_same_storage_numel_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_bin_edges.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_bin_edges_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_bin_edges_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_bin_edges_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_bin_edges_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_cts.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_cts_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_cts_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_cts_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_cts_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_tensors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_tensors_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_tensors_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_tensors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_histogramdd_from_bin_tensors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_index_put_impl_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_int_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_int_mm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_int_mm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_int_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_int_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_all_true.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_all_true_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_all_true_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_all_true_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_any_true.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_any_true_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_any_true_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_any_true_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_zerotensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_zerotensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_zerotensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_is_zerotensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_jagged_to_padded_dense_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_jagged_to_padded_dense_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_jagged_to_padded_dense_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_jagged_to_padded_dense_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lazy_clone.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lazy_clone_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lazy_clone_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lazy_clone_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_check_errors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_check_errors_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_check_errors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_check_errors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_det_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigvals.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigvals_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigvals_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigvals_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_eigvals_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_slogdet_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_solve_ex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_linalg_svd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_local_scalar_dense.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_local_scalar_dense_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_local_scalar_dense_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_local_scalar_dense_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_local_scalar_dense_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_backward_data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_log_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_logcumsumexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_logcumsumexp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_logcumsumexp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_logcumsumexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_logcumsumexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lstm_mps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lstm_mps_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lstm_mps_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lstm_mps_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lu_with_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lu_with_info_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lu_with_info_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_lu_with_info_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dep_token.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dep_token_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dep_token_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dep_token_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_dual_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_channel_quantized_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_channel_quantized_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_channel_quantized_tensor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_channel_quantized_tensor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_channel_quantized_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_channel_quantized_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_tensor_quantized_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_tensor_quantized_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_tensor_quantized_tensor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_tensor_quantized_tensor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_tensor_quantized_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_make_per_tensor_quantized_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_scale.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_scale_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_scale_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_scale_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_scale_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_masked_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mixed_dtypes_linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mixed_dtypes_linear_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mixed_dtypes_linear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mixed_dtypes_linear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_reshape.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_reshape_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_reshape_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_reshape_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_transpose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_transpose_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_transpose_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_transpose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mkldnn_transpose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_transpose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_transpose_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_transpose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_mps_convolution_transpose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_no_training.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_no_training_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_no_training_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_no_training_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_batch_norm_legit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_multi_head_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_multi_head_attention_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_multi_head_attention_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_multi_head_attention_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_multi_head_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_native_multi_head_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_neg_view_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_compute_contiguous_strides_offsets.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_compute_contiguous_strides_offsets_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_compute_contiguous_strides_offsets_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_compute_contiguous_strides_offsets_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_compute_contiguous_strides_offsets_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_and_nested_example.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_and_nested_example_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_and_nested_example_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_and_nested_example_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_from_padded_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_jagged_dummy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_jagged_dummy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_jagged_dummy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_lengths.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_lengths_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_lengths_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_offsets.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_offsets_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_offsets_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_ragged_idx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_ragged_idx_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_ragged_idx_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_get_values_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_select_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_select_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_select_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_sum_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_sum_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_sum_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_left_aligned.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_left_aligned_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_left_aligned_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_left_aligned_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_left_aligned_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_mask_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_tensor_list.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_tensor_list_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_tensor_list_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_from_tensor_list_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_size_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_softmax_with_shape.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_softmax_with_shape_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_softmax_with_shape_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_storage_offsets.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_storage_offsets_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_storage_offsets_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_storage_offsets_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_strides.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_strides_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_strides_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_tensor_strides_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_buffer_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nested_view_from_jagged_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_new_zeros_with_same_feature_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_new_zeros_with_same_feature_meta_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_new_zeros_with_same_feature_meta_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_new_zeros_with_same_feature_meta_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_available.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_available_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_available_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_available_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_spatial_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_spatial_convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_spatial_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnpack_spatial_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnz.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnz_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_nnz_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pack_padded_sequence_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_circular.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_circular_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_circular_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_circular_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_enum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_enum_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_enum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_enum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_packed_sequence.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_packed_sequence_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_packed_sequence_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pad_packed_sequence_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_padded_dense_to_jagged_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_padded_dense_to_jagged_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_padded_dense_to_jagged_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_padded_dense_to_jagged_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_forward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pdist_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pin_memory.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pin_memory_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pin_memory_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pin_memory_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_pin_memory_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_prelu_kernel_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_print.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_print_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_print_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_print_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_propagate_xla_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_propagate_xla_data_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_propagate_xla_data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_propagate_xla_data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_remove_batch_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_remove_batch_dim_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_remove_batch_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_remove_batch_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_alias_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_from_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_from_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_from_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_reshape_from_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_resize_output.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_resize_output_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_resize_output_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_resize_output_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_resize_output_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_rowwise_prune.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_rowwise_prune_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_rowwise_prune_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_rowwise_prune_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sample_dirichlet.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sample_dirichlet_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sample_dirichlet_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sample_dirichlet_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sample_dirichlet_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sample_dirichlet_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_saturate_weight_to_fp16.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_saturate_weight_to_fp16_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_saturate_weight_to_fp16_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_saturate_weight_to_fp16_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_attention_math.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_attention_math_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_attention_math_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_attention_math_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_cudnn_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_efficient_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_for_cpu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_dot_product_flash_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_mm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_scaled_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_segment_reduce_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_segment_reduce_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_segment_reduce_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_segment_reduce_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_segment_reduce_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_segment_reduce_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_shape_as_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_shape_as_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_shape_as_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_shape_as_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_slow_conv2d_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_draw.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_draw_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_draw_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_draw_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_ff.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_ff_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_ff_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_ff_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_initialize_state.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_initialize_state_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_initialize_state_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_initialize_state_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_scramble.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_scramble_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_scramble_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sobol_engine_scramble_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_backward_data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_addmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_addmm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_addmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_addmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_broadcast_to_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsc_tensor_unsafe.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsc_tensor_unsafe_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsc_tensor_unsafe_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsc_tensor_unsafe_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsr_tensor_unsafe.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsr_tensor_unsafe_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsr_tensor_unsafe_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_bsr_tensor_unsafe_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_unsafe.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_unsafe_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_unsafe_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_unsafe_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_with_dims.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_with_dims_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_with_dims_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_compressed_tensor_with_dims_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_unsafe.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_unsafe_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_unsafe_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_unsafe_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_and_tensors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_and_tensors_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_and_tensors_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_and_tensors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_and_tensors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_coo_tensor_with_dims_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csc_tensor_unsafe.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csc_tensor_unsafe_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csc_tensor_unsafe_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csc_tensor_unsafe_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_prod.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_prod_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_prod_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_prod_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_sum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_sum_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_sum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_sum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_tensor_unsafe.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_tensor_unsafe_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_tensor_unsafe_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_csr_tensor_unsafe_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_backward_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_backward_data_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_backward_data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_backward_data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_log_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mask_projection.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mask_projection_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mask_projection_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mask_projection_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_reduce_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_reduce_impl_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_reduce_impl_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_reduce_impl_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_reduce_impl_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_mm_reduce_impl_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_addmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_addmm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_addmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_addmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_dense.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_dense_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_dense_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_dense_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_apply_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_linear_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_linear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_linear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_mm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_tile.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_tile_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_tile_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_semi_structured_tile_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_backward_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_backward_data_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_backward_data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_backward_data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sparse_matmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sparse_matmul_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sparse_matmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sparse_matmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_sparse_sum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_spdiags.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_spdiags_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_spdiags_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_spdiags_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_spdiags_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_stack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_stack_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_stack_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_stack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_stack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_grad_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_grad_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_grad_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_grad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_grad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_standard_gamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_ambiguous_defaults.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_ambiguous_defaults_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_ambiguous_defaults_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_ambiguous_defaults_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_autograd_multiple_dispatch_view_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_check_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_check_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_check_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_check_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_functorch_fallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_functorch_fallback_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_functorch_fallback_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_functorch_fallback_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_functorch_fallback_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_filled_intlist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_filled_intlist_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_filled_intlist_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_filled_intlist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_filled_intlist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_floatlist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_floatlist_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_floatlist_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_floatlist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_floatlist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_intlist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_intlist_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_intlist_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_intlist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_optional_intlist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_parallel_materialize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_parallel_materialize_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_parallel_materialize_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_parallel_materialize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_serialization_subcmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_serialization_subcmul_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_serialization_subcmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_serialization_subcmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_string_default.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_string_default_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_string_default_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_string_default_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_warn_in_autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_warn_in_autograd_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_warn_in_autograd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_test_warn_in_autograd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_gru_cell_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_gru_cell_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_gru_cell_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_gru_cell_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_lstm_cell_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_lstm_cell_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_lstm_cell_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_differentiable_lstm_cell_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_gru_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_impl_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_impl_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_impl_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_impl_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_thnn_fused_lstm_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_cpu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_cpu_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_cpu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_cpu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_dense.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_dense_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_dense_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_dense_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsc_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsc_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsr_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_bsr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csc_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csc_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csr_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_csr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_semi_structured.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_semi_structured_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_semi_structured_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_to_sparse_semi_structured_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transform_bias_rescale_qkv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transform_bias_rescale_qkv_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transform_bias_rescale_qkv_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transform_bias_rescale_qkv_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transform_bias_rescale_qkv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transform_bias_rescale_qkv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transformer_encoder_layer_fwd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transformer_encoder_layer_fwd_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transformer_encoder_layer_fwd_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transformer_encoder_layer_fwd_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transformer_encoder_layer_fwd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_transformer_encoder_layer_fwd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_trilinear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_trilinear_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_trilinear_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_trilinear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_trilinear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_multi_head_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_multi_head_attention_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_multi_head_attention_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_multi_head_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_multi_head_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_scaled_dot_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_scaled_dot_attention_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_scaled_dot_attention_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_scaled_dot_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_triton_scaled_dot_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique2_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique2_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique2_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unique_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unpack_dual.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unpack_dual_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unpack_dual_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unpack_dual_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_put.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_put_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_put_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_index_put_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_view.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_view_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_view_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_unsafe_view_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bicubic2d_aa_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_bilinear2d_aa_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_upsample_nearest_exact3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_ctc_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_ctc_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_ctc_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_ctc_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_rnn_flatten_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_rnn_flatten_weight_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_rnn_flatten_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_use_cudnn_rnn_flatten_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_compressed_sparse_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_compressed_sparse_indices_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_compressed_sparse_indices_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_compressed_sparse_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_compressed_sparse_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsc_tensor_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsc_tensor_args_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsc_tensor_args_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsc_tensor_args_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsr_tensor_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsr_tensor_args_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsr_tensor_args_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_bsr_tensor_args_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_compressed_tensor_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_compressed_tensor_args_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_compressed_tensor_args_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_compressed_tensor_args_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_coo_tensor_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_coo_tensor_args_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_coo_tensor_args_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_coo_tensor_args_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csc_tensor_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csc_tensor_args_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csc_tensor_args_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csc_tensor_args_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csr_tensor_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csr_tensor_args_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csr_tensor_args_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_validate_sparse_csr_tensor_args_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_values_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_version.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_version_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_version_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_version_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int4pack_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int4pack_mm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int4pack_mm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int4pack_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int4pack_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int8pack_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int8pack_mm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int8pack_mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_int8pack_mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_differentiable_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_differentiable_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_differentiable_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_differentiable_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_interface_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/_weight_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/abs.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/abs_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/abs_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/abs_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/abs_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/abs_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/absolute.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/absolute_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/absolute_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/absolute_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acos_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/acosh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_avg_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adaptive_max_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/add_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addbmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addbmm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addbmm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addbmm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addbmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addbmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcdiv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addcmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addmv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addr_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/addr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adjoint.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adjoint_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adjoint_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/adjoint_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/affine_grid_generator_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alias_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_as.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_as_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_as_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_as_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_tensors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_tensors_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_tensors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_tensors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_to.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_to_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_to_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/align_to_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/all_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/allclose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/allclose_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/allclose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/allclose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alpha_dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alpha_dropout_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alpha_dropout_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/alpha_dropout_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/amin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/aminmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/and.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/and_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/and_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/and_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/angle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/angle_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/angle_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/angle_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/angle_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/any_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arange_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccos.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccos_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccos_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccos_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccosh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccosh_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccosh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arccosh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsin_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsinh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsinh_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsinh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arcsinh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctan_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctanh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctanh_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctanh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/arctanh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argmin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argsort_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argwhere.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argwhere_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argwhere_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/argwhere_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_scatter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_scatter_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_scatter_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_scatter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/as_strided_scatter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/asinh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atan_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atanh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/atleast_3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/avg_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/baddbmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bartlett_window.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bartlett_window_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bartlett_window_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bartlett_window_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_elemt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_elemt_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_elemt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_elemt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_elemt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_reduce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_reduce_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_reduce_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_reduce_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_backward_reduce_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_elemt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_elemt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_elemt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_elemt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_with_counts.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_with_counts_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_with_counts_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_with_counts_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_gather_stats_with_counts_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_stats.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_stats_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_stats_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_stats_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_stats_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_update_stats.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_update_stats_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_update_stats_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_update_stats_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_update_stats_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/batch_norm_update_stats_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bernoulli_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bilinear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bilinear_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bilinear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bilinear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_with_logits.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_with_logits_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_with_logits_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binary_cross_entropy_with_logits_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bincount.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bincount_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bincount_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bincount_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bincount_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bincount_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binomial.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binomial_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binomial_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binomial_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binomial_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/binomial_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_and_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_left_shift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_not_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_or_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_right_shift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bitwise_xor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/blackman_window.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/blackman_window_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/blackman_window_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/blackman_window_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/block_diag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/block_diag_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/block_diag_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/block_diag_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_tensors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_tensors_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_tensors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_tensors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_to.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_to_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_to_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/broadcast_to_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bucketize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bucketize_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bucketize_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bucketize_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bucketize_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/bucketize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/can_cast.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/can_cast_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/can_cast_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/can_cast_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cartesian_prod.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cartesian_prod_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cartesian_prod_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cartesian_prod_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cat_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cauchy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ccol_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cdist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cdist_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cdist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cdist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ceil_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/celu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/celu_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/celu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/celu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chain_matmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chain_matmul_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chain_matmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chain_matmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chalf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chalf_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chalf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chalf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/channel_shuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/channel_shuffle_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/channel_shuffle_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/channel_shuffle_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/channel_shuffle_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/channel_shuffle_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_inverse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_inverse_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_inverse_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_inverse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_inverse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_solve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_solve_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_solve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cholesky_solve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/choose_qparams_optimized.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/choose_qparams_optimized_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/choose_qparams_optimized_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/choose_qparams_optimized_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chunk.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chunk_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chunk_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/chunk_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_max_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_min_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clamp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clip.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clip_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clip_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clip_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clone.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clone_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clone_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/clone_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/coalesce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/coalesce_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/coalesce_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/coalesce_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col2im.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col2im_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col2im_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col2im_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col2im_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/col_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/column_stack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/column_stack_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/column_stack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/column_stack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/combinations.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/combinations_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/combinations_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/combinations_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/complex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/complex_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/complex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/complex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/complex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/complex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concat_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concat_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concat_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concatenate.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concatenate_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concatenate_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/concatenate_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conj_physical_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/constant_pad_nd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/constant_pad_nd_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/constant_pad_nd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/constant_pad_nd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/contiguous.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/contiguous_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/contiguous_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/contiguous_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_depthwise3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_depthwise3d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_depthwise3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_depthwise3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_depthwise3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_tbc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/conv_transpose3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_overrideable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_overrideable_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_overrideable_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_backward_overrideable_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_overrideable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_overrideable_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_overrideable_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/convolution_overrideable_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_sparse_to_sparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_sparse_to_sparse_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_sparse_to_sparse_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_sparse_to_sparse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copy_sparse_to_sparse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/copysign_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/corrcoef.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/corrcoef_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/corrcoef_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/corrcoef_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cos_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_embedding_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_embedding_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_embedding_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_embedding_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_similarity.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_similarity_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_similarity_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cosine_similarity_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/count_nonzero.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/count_nonzero_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/count_nonzero_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/count_nonzero_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/count_nonzero_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/count_nonzero_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cov.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cov_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cov_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cov_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_entropy_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_entropy_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_entropy_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_entropy_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cross_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/crow_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ctc_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ctc_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ctc_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ctc_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_affine_grid_generator_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_batch_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_add_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_add_relu_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_add_relu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_add_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_add_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_relu_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_relu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_transpose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_transpose_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_transpose_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_transpose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_convolution_transpose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_grid_sampler_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_is_acceptable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_is_acceptable_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_is_acceptable_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cudnn_is_acceptable_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummaxmin_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummaxmin_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummaxmin_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummaxmin_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummin_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummin_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cummin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumprod_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumsum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumulative_trapezoid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumulative_trapezoid_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumulative_trapezoid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/cumulative_trapezoid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/data_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/deg2rad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/deg2rad_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/deg2rad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/deg2rad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dense_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dense_dim_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dense_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dense_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dequantize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dequantize_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dequantize_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dequantize_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dequantize_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dequantize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/det.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/det_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/det_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/det_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/detach_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_embed.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_embed_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_embed_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_embed_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_embed_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diag_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagflat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagflat_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagflat_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagflat_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_scatter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_scatter_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_scatter_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_scatter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diagonal_scatter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diff.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diff_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diff_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/diff_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/digamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dist_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/div_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/divide.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/divide_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/divide_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/divide_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dot_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dot_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dot_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dropout_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dropout_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dropout_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dsplit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dsplit_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dsplit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dsplit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dstack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dstack_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dstack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/dstack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/einsum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/einsum_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/einsum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/einsum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/elu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_bag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_bag_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_bag_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_bag_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_dense_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_dense_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_dense_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_dense_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_dense_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_dense_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_renorm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_sparse_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_sparse_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_sparse_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/embedding_sparse_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_permuted.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_permuted_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_permuted_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_permuted_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_quantized.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_quantized_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_quantized_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_quantized_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/empty_strided_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eq_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/equal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/equal_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/equal_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/equal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/equal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/erfinv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_as.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_as_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_as_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_as_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expand_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/expm1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/exponential_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/eye_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_cachemask_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_channel_affine_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_cachemask_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fake_quantize_per_tensor_affine_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_fp32_activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_fp32_activation_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_fp32_activation_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_fp32_activation_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_fp16_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_fp32_activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_fp32_activation_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_fp32_activation_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_fp32_activation_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_int8_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_quantize_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_quantize_weight_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_quantize_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_linear_quantize_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_gemm_matrix_fp16.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_gemm_matrix_fp16_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_gemm_matrix_fp16_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_gemm_matrix_fp16_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_quantized_matrix.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_quantized_matrix_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_quantized_matrix_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fbgemm_pack_quantized_matrix_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_alpha_dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_alpha_dropout_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_alpha_dropout_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_alpha_dropout_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_dropout_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_dropout_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/feature_dropout_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftfreq.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftfreq_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftfreq_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftfreq_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftshift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftshift_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftshift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_fftshift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfftn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfftn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfftn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_hfftn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftshift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftshift_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftshift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ifftshift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfftn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfftn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfftn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_ihfftn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfftn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfftn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfftn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_irfftn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftfreq.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftfreq_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftfreq_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftfreq_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fft_rfftn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_diagonal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_diagonal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_diagonal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_diagonal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fill_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fix.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fix_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fix_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fix_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_dense_tensors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_dense_tensors_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_dense_tensors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_dense_tensors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flatten_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flip.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flip_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flip_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flip_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flip_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flip_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fliplr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fliplr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fliplr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fliplr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flipud.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flipud_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flipud_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/flipud_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/float_power.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/float_power_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/float_power_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/float_power_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_divide_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/floor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fmod_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frac_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fractional_max_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frexp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frexp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frexp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frobenius_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frobenius_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frobenius_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/frobenius_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/from_file.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/from_file_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/from_file_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/from_file_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/from_file_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/full_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fused_moving_avg_obs_fake_quant.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fused_moving_avg_obs_fake_quant_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fused_moving_avg_obs_fake_quant_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/fused_moving_avg_obs_fake_quant_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gather_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gcd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ge_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gelu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geometric_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geqrf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geqrf_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geqrf_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geqrf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/geqrf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ger.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ger_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ger_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ger_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_jvp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_jvp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_jvp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_jvp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_jvp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_jvp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_jvp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_jvp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_jvp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_jvp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_jvp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_jvp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/glu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gradient.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gradient_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gradient_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gradient_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_equal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_equal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_equal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_equal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/greater_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/grid_sampler_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/group_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/group_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/group_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/group_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gru_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/gt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hamming_window.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hamming_window_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hamming_window_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hamming_window_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hann_window.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hann_window_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hann_window_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hann_window_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardshrink_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardsigmoid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardswish_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hardtanh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/heaviside_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hinge_embedding_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hinge_embedding_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hinge_embedding_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hinge_embedding_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histc_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogram.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogram_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogram_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogram_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogramdd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogramdd_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogramdd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/histogramdd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hsplit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hsplit_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hsplit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hsplit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hspmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hspmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hspmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hstack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hstack_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hstack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hstack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/huber_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/hypot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/i0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/igammac_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/im2col.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/im2col_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/im2col_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/im2col_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/im2col_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/imag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/imag_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/imag_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/imag_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_add_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_fill_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_put.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_put_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_put_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_put_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_reduce_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/index_select_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/infinitely_differentiable_gelu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/infinitely_differentiable_gelu_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/infinitely_differentiable_gelu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/infinitely_differentiable_gelu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inner.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inner_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inner_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inner_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/instance_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/instance_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/instance_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/instance_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/int_repr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/int_repr_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/int_repr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/int_repr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inverse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inverse_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inverse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/inverse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_coalesced.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_coalesced_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_coalesced_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_coalesced_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_complex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_complex_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_complex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_complex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_conj.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_conj_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_conj_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_conj_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_distributed.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_distributed_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_distributed_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_distributed_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_floating_point.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_floating_point_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_floating_point_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_floating_point_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_inference.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_inference_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_inference_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_inference_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_leaf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_leaf_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_leaf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_leaf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_neg.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_neg_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_neg_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_neg_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_nonzero.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_nonzero_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_nonzero_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_nonzero_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_pinned.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_pinned_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_pinned_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_pinned_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_pinned_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_same_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_same_size_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_same_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_same_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_set_to.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_set_to_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_set_to_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_set_to_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_set_to_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_signed.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_signed_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_signed_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_signed_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_vulkan_available.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_vulkan_available_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_vulkan_available_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/is_vulkan_available_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isclose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isclose_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isclose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isclose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isfinite.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isfinite_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isfinite_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isfinite_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isinf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isinf_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isinf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isinf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isnan.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isnan_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isnan_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isnan_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isnan_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isnan_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isneginf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isposinf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isreal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isreal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isreal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/isreal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/istft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/istft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/istft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/istft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/item.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/item_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/item_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/item_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kaiser_window.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kaiser_window_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kaiser_window_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kaiser_window_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kl_div.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kl_div_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kl_div_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kl_div_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kron.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kron_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kron_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kron_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/kthvalue_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/l1_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/l1_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/l1_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/l1_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/layer_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/layer_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/layer_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/layer_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lcm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ldexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ldexp_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ldexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ldexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/le_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/leaky_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lerp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_equal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_equal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_equal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_equal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/less_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lgamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_fresh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cholesky_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cond.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cond_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cond_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cond_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_cross_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_det.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_det_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_det_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_det_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_diagonal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_diagonal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_diagonal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_diagonal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eig.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eig_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eig_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eig_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eig_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigh_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvals.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvals_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvals_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvals_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvals_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvals_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvalsh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvalsh_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvalsh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_eigvalsh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_householder_product.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_householder_product_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_householder_product_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_householder_product_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_householder_product_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_inv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_factor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_ldl_solve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lstsq.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lstsq_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lstsq_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lstsq_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lstsq_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lstsq_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_factor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_lu_solve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matmul_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_exp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_exp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_exp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_exp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_exp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_exp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_power.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_power_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_power_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_power_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_rank.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_rank_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_rank_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_matrix_rank_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_multi_dot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_multi_dot_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_multi_dot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_multi_dot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_pinv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_pinv_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_pinv_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_pinv_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_pinv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_pinv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_qr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_slogdet.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_slogdet_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_slogdet_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_slogdet_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_ex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_ex_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_ex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_ex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_triangular.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_triangular_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_triangular_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_triangular_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_solve_triangular_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svd_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svdvals.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svdvals_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svdvals_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_svdvals_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorinv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorinv_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorinv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorinv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorsolve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorsolve_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorsolve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_tensorsolve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vander.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vander_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vander_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vander_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vecdot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vecdot_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vecdot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vecdot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linalg_vector_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/linspace_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log10_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log1p_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_normal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_sigmoid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_softmax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_softmax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/log_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logaddexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logcumsumexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logcumsumexp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logcumsumexp_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logcumsumexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logcumsumexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logdet.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logdet_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logdet_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logdet_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_and.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_and_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_and_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_and_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_and_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_and_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_not.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_not_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_not_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_not_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_not_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_not_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_or.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_or_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_or_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_or_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_or_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_or_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_xor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_xor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_xor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_xor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_xor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logical_xor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logspace_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logsumexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logsumexp_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logsumexp_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logsumexp_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logsumexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/logsumexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lshift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_mps_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_mps_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_mps_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_mps_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lstm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_solve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_solve_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_solve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_solve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/lu_unpack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mH.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mH_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mH_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mH_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mT.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mT_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mT_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mT_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/margin_ranking_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/margin_ranking_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/margin_ranking_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/margin_ranking_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_fill_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_scatter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/masked_select_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matmul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_H.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_H_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_H_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_H_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_exp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_power.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_power_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_power_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/matrix_power_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_with_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_with_indices_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_with_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool1d_with_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool2d_with_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_pool3d_with_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/max_unpool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/maximum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mean_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/median_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/meshgrid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/meshgrid_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/meshgrid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/meshgrid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/min_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/minimum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_batch_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_add_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_add_relu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_add_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_add_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_relu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_transpose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_transpose_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_transpose_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_transpose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_convolution_transpose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_depthwise_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_depthwise_convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_depthwise_convolution_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_depthwise_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_depthwise_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/miopen_rnn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mish_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_adaptive_avg_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_convolution.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_convolution_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_convolution_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_convolution_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_input.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_input_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_input_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_input_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_weights.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_weights_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_weights_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_backward_weights_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_linear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_max_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv2d_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv2d_weight_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv2d_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv2d_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv3d_weight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv3d_weight_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv3d_weight_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_reorder_conv3d_weight_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mkldnn_rnn_layer_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mode_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/moveaxis.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/moveaxis_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/moveaxis_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/moveaxis_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/movedim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/movedim_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/movedim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/movedim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_transpose_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_transpose_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_transpose_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mps_convolution_transpose_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mse_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/msort.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/msort_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/msort_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/msort_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mul_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multi_margin_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multilabel_margin_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multinomial.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multinomial_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multinomial_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multinomial_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multinomial_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multiply.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multiply_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multiply_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/multiply_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mv_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mvlgamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mvlgamma_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mvlgamma_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mvlgamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mvlgamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/mvlgamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nan_to_num.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nan_to_num_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nan_to_num_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nan_to_num_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nan_to_num_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nan_to_num_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmean.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmean_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmean_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmean_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanmedian_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanquantile.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanquantile_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanquantile_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nanquantile_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nansum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nansum_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nansum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nansum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nansum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_copy_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/narrow_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_batch_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_channel_shuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_channel_shuffle_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_channel_shuffle_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_channel_shuffle_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_channel_shuffle_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_dropout_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_group_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_layer_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/native_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ne_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/neg_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/negative.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/negative_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/negative_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/negative_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nested_to_padded_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nested_to_padded_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nested_to_padded_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nested_to_padded_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_strided.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_strided_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_strided_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_strided_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_empty_strided_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_full.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_full_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_full_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_full_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_ones.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_ones_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_ones_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_ones_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_zeros.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_zeros_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_zeros_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/new_zeros_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nextafter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_nd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_nd_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_nd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_nd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nll_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_numpy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_numpy_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_numpy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_numpy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_static.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_static_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_static_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nonzero_static_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_except_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_except_dim_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_except_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_except_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/normal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/not_equal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/not_equal_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/not_equal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/not_equal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nuclear_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nuclear_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nuclear_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/nuclear_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/numpy_T.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/numpy_T_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/numpy_T_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/numpy_T_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/one_hot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/one_hot_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/one_hot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/one_hot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ones_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/or.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/or_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/or_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/or_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/orgqr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/orgqr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/orgqr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/orgqr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ormqr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ormqr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ormqr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ormqr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ormqr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/outer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/outer_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/outer_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/outer_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/output_nr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/output_nr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/output_nr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/output_nr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_sequence.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_sequence_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_sequence_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pad_sequence_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pairwise_distance.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pairwise_distance_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pairwise_distance_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pairwise_distance_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pdist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pdist_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pdist_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pdist_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/permute_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pin_memory.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pin_memory_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pin_memory_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pin_memory_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pinverse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pinverse_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pinverse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pinverse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_shuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_shuffle_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_shuffle_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_shuffle_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_shuffle_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_shuffle_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_unshuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_unshuffle_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_unshuffle_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_unshuffle_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_unshuffle_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pixel_unshuffle_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_nll_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_nll_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_nll_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_nll_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/poisson_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polar.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polar_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polar_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polar_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polar_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polar_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/polygamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/positive.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/positive_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/positive_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/positive_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/pow_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prelu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prelu_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prelu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prelu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/prod_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/promote_types.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/promote_types_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/promote_types_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/promote_types_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/put_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_axis.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_axis_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_axis_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_scales.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_scales_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_scales_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_scales_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_zero_points.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_zero_points_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_zero_points_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_per_channel_zero_points_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_scale.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_scale_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_scale_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_zero_point.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_zero_point_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/q_zero_point_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qscheme.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qscheme_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/qscheme_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantile.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantile_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantile_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantile_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_channel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_channel_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_channel_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_channel_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_channel_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_channel_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_dynamic.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_dynamic_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_dynamic_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_dynamic_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_dynamic_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_dynamic_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantize_per_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_batch_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_batch_norm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_batch_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_batch_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_gru_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_gru_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_gru_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_gru_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_lstm_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_lstm_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_lstm_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_lstm_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool1d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool2d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool3d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_max_pool3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_relu_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_relu_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_relu_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_relu_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_tanh_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_tanh_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_tanh_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/quantized_rnn_tanh_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rad2deg.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rad2deg_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rad2deg_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rad2deg_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rand_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randint_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_like_compositeimplicitautogradnestedtensor_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/random_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randperm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randperm_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randperm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randperm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randperm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/randperm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/range_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ravel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ravel_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ravel_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/ravel_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/real.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/real_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/real_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/real_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reciprocal_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/record_stream.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/record_stream_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/record_stream_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/record_stream_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/refine_names.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/refine_names_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/refine_names_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/refine_names_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reflection_pad3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu6.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu6_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu6_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu6_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/remainder_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rename.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rename_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rename_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rename_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/renorm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_interleave_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/repeat_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/replication_pad3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/requires_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/requires_grad_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/requires_grad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/requires_grad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_as.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_as_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_as_compositeimplicitautogradnestedtensor_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_as_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_as_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_compositeimplicitautogradnestedtensor_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/reshape_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_sparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_sparse_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_sparse_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_sparse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_as_sparse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_conj.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_conj_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_conj_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_conj_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_neg.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_neg_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_neg_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/resolve_neg_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/result_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/result_type_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/result_type_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/result_type_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retain_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retain_grad_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retain_grad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retain_grad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retains_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retains_grad_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retains_grad_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/retains_grad_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rms_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rms_norm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rms_norm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rms_norm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_relu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_cell.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_cell_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_cell_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_cell_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rnn_tanh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/roll.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/roll_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/roll_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/roll_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/roll_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/roll_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rot90.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rot90_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rot90_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rot90_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/round_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_stack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_stack_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_stack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/row_stack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rrelu_with_noise_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rshift_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsqrt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsub.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsub_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsub_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsub_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsub_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/rsub_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scalar_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scalar_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scalar_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scalar_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scaled_dot_product_attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scaled_dot_product_attention_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scaled_dot_product_attention_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scaled_dot_product_attention_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_add_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/scatter_reduce_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/searchsorted.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/searchsorted_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/searchsorted_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/searchsorted_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/searchsorted_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/segment_reduce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/segment_reduce_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/segment_reduce_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/segment_reduce_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/segment_reduce_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/segment_reduce_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_scatter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_scatter_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_scatter_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_scatter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/select_scatter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/selu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/selu_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/selu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/selu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_data_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_data_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_data_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/set_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sgn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sigmoid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sign_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/signbit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/silu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sin_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sinh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/size_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_inverse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_inverse_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_inverse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_inverse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_scatter.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_scatter_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_scatter_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_scatter_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slice_scatter_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slogdet.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slogdet_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slogdet_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slogdet_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_forward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_forward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_forward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_forward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated2d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated3d_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_dilated3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/slow_conv_transpose3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/smooth_l1_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/soft_margin_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softmax_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softmax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softplus_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/softshrink_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sort_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsc_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsc_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsc_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsc_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsr_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsr_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsr_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_bsr_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_compressed_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_compressed_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_compressed_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_compressed_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_coo_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_coo_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_coo_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_coo_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_coo_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csc_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csc_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csc_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csc_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csr_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csr_tensor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csr_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_csr_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_dim_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_mask.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_mask_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_mask_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_mask_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_and_clear.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_and_clear_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_and_clear_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_and_clear_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_and_clear_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_resize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_sampled_addmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_sampled_addmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sparse_sampled_addmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_airy_ai_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_j1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_bessel_y1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_t_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_u_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_v_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_chebyshev_polynomial_w_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_digamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_digamma_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_digamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_digamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_entr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erf.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erf_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erf_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erf_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfc_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfcx_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfinv.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfinv_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfinv_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_erfinv_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_exp2.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_exp2_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_exp2_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_exp2_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expit_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expm1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expm1_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expm1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_expm1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammainc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammainc_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammainc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammainc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaincc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaincc_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaincc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaincc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaln.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaln_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaln_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_gammaln_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_h_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_hermite_polynomial_he_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i0e_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_i1e_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_laguerre_polynomial_l_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_legendre_polynomial_p_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log1p.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log1p_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log1p_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log1p_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_ndtr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_softmax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_log_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logit_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logsumexp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logsumexp_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logsumexp_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_logsumexp_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_i1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_modified_bessel_k1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_multigammaln.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_multigammaln_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_multigammaln_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_multigammaln_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_ndtri_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_polygamma.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_polygamma_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_polygamma_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_polygamma_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_psi.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_psi_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_psi_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_psi_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_round.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_round_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_round_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_round_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_scaled_modified_bessel_k1_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_t_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_u_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_v_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_shifted_chebyshev_polynomial_w_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_sinc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_sinc_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_sinc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_sinc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_softmax_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_softmax_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_softmax_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_spherical_bessel_j0_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlog1py_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlogy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlogy_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlogy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_xlogy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/special_zeta_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_copy_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/split_with_sizes_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sqrt_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/square.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/square_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/square_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/square_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/squeeze_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sspaddmm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sspaddmm_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sspaddmm_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sspaddmm_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sspaddmm_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sspaddmm_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stack_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_mean_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/std_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stft.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stft_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stft_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stft_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stride.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stride_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stride_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/stride_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sub_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/subtract.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/subtract_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/subtract_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/subtract_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_to_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_to_size_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_to_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sum_to_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/svd.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/svd_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/svd_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/svd_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapaxes.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapaxes_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapaxes_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapaxes_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapdims.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapdims_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapdims_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/swapdims_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_for_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_for_size_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_for_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_for_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_constrain_range_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_numel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_numel_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_numel_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_numel_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_size.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_size_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_size_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_size_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_storage_offset.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_storage_offset_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_storage_offset_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_storage_offset_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_stride.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_stride_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_stride_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/sym_stride_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/t_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_along_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_along_dim_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_along_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_along_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/take_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tan_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tanh_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensor_split.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensor_split_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensor_split_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensor_split_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensordot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensordot_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensordot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tensordot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/thnn_conv2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/thnn_conv2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/thnn_conv2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/thnn_conv2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/threshold_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tile.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tile_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tile_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tile_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_dense_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_mkldnn_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_padded_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_padded_tensor_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_padded_tensor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_padded_tensor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsc_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_bsr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csc_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csr.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csr_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csr_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_csr_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/to_sparse_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/topk_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trace_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/transpose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapezoid.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapezoid_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapezoid_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapezoid_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapz.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapz_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapz_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trapz_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triangular_solve_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_indices_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_indices_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/tril_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triplet_margin_loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triplet_margin_loss_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triplet_margin_loss_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triplet_margin_loss_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_indices.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_indices_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_indices_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_indices_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_indices_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_indices_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/triu_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/true_divide.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/true_divide_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/true_divide_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/true_divide_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/trunc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/type_as.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/type_as_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/type_as_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/type_as_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unbind_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_dense_tensors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_dense_tensors_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_dense_tensors_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_dense_tensors_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unflatten_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_backward_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unfold_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/uniform_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_consecutive.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_consecutive_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_consecutive_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_consecutive_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_consecutive_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_consecutive_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_consecutive.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_consecutive_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_consecutive_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_consecutive_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_consecutive_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_consecutive_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unique_dim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_chunk.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_chunk_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_chunk_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_chunk_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_with_sizes.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_with_sizes_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_with_sizes_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsafe_split_with_sizes_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/unsqueeze_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bicubic2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_bilinear2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_linear1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest1d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest2d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_nearest3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/upsample_trilinear3d_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/value_selecting_reduction_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/value_selecting_reduction_backward_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/value_selecting_reduction_backward_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/value_selecting_reduction_backward_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/values_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vander.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vander_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vander_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vander_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_mean_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/var_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vdot.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vdot_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vdot_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vdot_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vdot_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vdot_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_complex_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_as_real_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_copy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_copy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_copy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_copy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/view_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vsplit.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vsplit_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vsplit_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vsplit_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vstack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vstack_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vstack_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/vstack_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/where.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/where_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/where_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/where_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/where_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/where_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_compositeexplicitautogradnonfunctional_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_meta.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xlogy_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xor_compositeimplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xor_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/xor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero_cpu_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero_cuda_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero_meta_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zero_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_like.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_like_compositeexplicitautograd_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_like_compositeimplicitautogradnestedtensor_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_like_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_like_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_native.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops copying torch/include/ATen/ops/zeros_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/ops creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/ApplyGridUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/AsmUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/Atomic.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/DeviceUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPApplyUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPGraphsUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPTensorMethods.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/NumericLimits.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/PhiloxUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/ScanUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/cub.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/cub_definitions.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/ATenHIPGeneral.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/CachingHostAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/EmptyTensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/Exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPBlas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPConfig.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPContext.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPContextLight.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPDataType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPDevice.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPEvent.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPGeneratorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPGraph.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPSparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPSparseBlas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPSparseDescriptors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/HIPUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/PeerToPeerAccess.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/PhiloxCudaState.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/PinnedMemoryAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/Sleep.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/ThrustAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/cub.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/jiterator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/jiterator_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip copying torch/include/ATen/hip/llvm_jit_strings.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/IndexUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/IntegerDivider.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/OffsetCalculator.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/PhiloxCudaStateRaw.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/TensorInfo.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/UnpackRaw.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/DeviceThreadHandles.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/HIPHooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/KernelUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail copying torch/include/ATen/hip/detail/LazyNVRTC.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/detail creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/impl copying torch/include/ATen/hip/impl/HIPAllocatorMasqueradingAsCUDA.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/impl copying torch/include/ATen/hip/impl/HIPCachingAllocatorMasqueradingAsCUDA.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/impl copying torch/include/ATen/hip/impl/HIPGuardImplMasqueradingAsCUDA.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/impl copying torch/include/ATen/hip/impl/HIPStreamMasqueradingAsCUDA.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/impl creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/GemmCommon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/GemmHipblaslt.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/GemmRocblas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/StreamTimer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/Tunable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/TunableGemm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable copying torch/include/ATen/hip/tunable/TunableOp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/hip/tunable creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/EmptyTensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/IndexKernels.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSAllocatorInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSDevice.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSEvent.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSGeneratorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSGuardImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSHooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSProfiler.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps copying torch/include/ATen/mps/MPSStream.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/mps creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen copying torch/include/ATen/miopen/Descriptors.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen copying torch/include/ATen/miopen/Exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen copying torch/include/ATen/miopen/Handle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen copying torch/include/ATen/miopen/Types.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen copying torch/include/ATen/miopen/Utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen copying torch/include/ATen/miopen/miopen-wrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/miopen creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/AcceleratorHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/CUDAHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/FunctionTraits.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/HIPHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/IPUHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/MAIAHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/MPSHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/MTIAHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/PrivateUse1HooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail copying torch/include/ATen/detail/XPUHooksInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/detail creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/AdaptivePooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/AmpKernels.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/BatchLinearAlgebra.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/BinaryOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/BucketizationUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/CPUBlas.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/CPUFallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/CanUse32BitIndexMath.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ComplexHelper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/CompositeRandomAccessor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/CompositeRandomAccessorCommon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ConvUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ConvolutionMM3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Cross.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/DilatedConvolutionUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/DispatchStub.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Distance.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/DistributionTemplates.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Distributions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/EmbeddingBag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Fill.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ForeachUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/FractionalMaxPooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/FunctionOfAMatrixUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/FusedAdagrad.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/FusedAdam.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/FusedSGD.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/GridSampler.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/GridSamplerUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Histogram.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/IndexKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/IndexingUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Lerp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/LinearAlgebra.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/LinearAlgebraUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/LossMulti.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Math.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/MathBitFallThroughLists.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/MathBitsFallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/MaxPooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/NonEmptyUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/NonSymbolicBC.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Normalization.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Padding.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/PixelShuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/PointwiseOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Pool.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Pow.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/RNN.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/RangeFactories.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ReduceAllOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ReduceOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ReduceOpsUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ReductionType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Repeat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Resize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ResizeCommon.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/ScatterGatherChecks.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/SegmentReduce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/SharedReduceOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/SobolEngineOpsUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Sorting.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/SortingUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/SparseTensorUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/SpectralOpsUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/StridedRandomAccessor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorAdvancedIndexing.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorAdvancedIndexingUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorCompare.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorConversions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorDimApply.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorFactories.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorIterator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorIteratorDynamicCasting.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorProperties.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorShape.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TensorTransformations.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TopKImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TransposeType.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TriangularOpsUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/TypeProperties.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/UnaryOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Unfold2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/Unfold3d.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/UnfoldBackward.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/UpSample.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/batch_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/group_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/im2col.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/im2col_shape_check.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/layer_norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/verbose_wrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native copying torch/include/ATen/native/vol2col.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/AtomicAddFloat.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/CatKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/ChannelShuffleKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/CopyKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/DepthwiseConvKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/DistributionTemplates.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/GridSamplerKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/IndexKernelUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/Intrinsics.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/IsContiguous.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/LogAddExp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/Loops.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/MaxUnpoolKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/PixelShuffleKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/Reduce.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/ReduceUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/SampledAddmmKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/SerialStackImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/SoftmaxKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/SpmmReduceKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/StackKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/UpSampleKernelAVXAntialias.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/WeightNormKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/avx_mathfun.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/int_mm_kernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/mixed_data_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/moments_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu copying torch/include/ATen/native/cpu/zmath.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cpu creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/BinaryInternal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/CompositeRandomAccessor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/CuFFTPlanCache.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/CuFFTUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/DistributionTemplates.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Distributions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/GridSampler.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/IndexKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/LaunchUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/MiscUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/ReduceOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Resize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/ScanKernels.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Sort.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/SortStable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Sorting.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/TensorModeKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/TensorTopK.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/jit_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/thread_constants.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/CUDAJitLoops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/CUDALoops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/DeviceSqrt.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/EmbeddingBackwardKernel.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/ForeachFunctors.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/ForeachMinMaxFunctors.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/GridSampler.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/JitLoops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/KernelUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Loops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Math.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/MemoryAccess.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/MultiTensorApply.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Normalization.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/PersistentSoftmax.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Pow.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Randperm.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/Reduce.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/ScanUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/SortUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/SortingCommon.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/SortingRadixSelect.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/TensorModeKernel.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/UniqueCub.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/UpSample.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/block_reduce.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/fused_adam_amsgrad_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/fused_adam_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/fused_adam_utils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/fused_adamw_amsgrad_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/fused_adamw_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/im2col.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/reduction_template.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda copying torch/include/ATen/native/cuda/vol2col.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/cuda creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/BinaryInternal.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/CompositeRandomAccessor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/CuFFTPlanCache.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/CuFFTUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/DistributionTemplates.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Distributions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/GridSampler.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/IndexKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/LaunchUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/MiscUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/ReduceOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Resize.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/ScanKernels.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Sort.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/SortStable.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Sorting.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/TensorModeKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/TensorTopK.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/jit_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/thread_constants.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/DeviceSqrt.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/EmbeddingBackwardKernel.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/ForeachFunctors.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/ForeachMinMaxFunctors.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/GridSampler.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/HIPJitLoops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/HIPLoops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/JitLoops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/KernelUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Loops.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Math.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/MemoryAccess.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/MultiTensorApply.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Normalization.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/PersistentSoftmax.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Pow.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Randperm.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/Reduce.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/ScanUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/SortUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/SortingCommon.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/SortingRadixSelect.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/TensorModeKernel.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/UniqueCub.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/UpSample.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/block_reduce.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/fused_adam_amsgrad_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/fused_adam_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/fused_adam_utils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/fused_adamw_amsgrad_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/fused_adamw_impl.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/im2col.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/reduction_template.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip copying torch/include/ATen/native/hip/vol2col.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/hip creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps copying torch/include/ATen/native/mps/Copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps copying torch/include/ATen/native/mps/MPSGraphSonomaOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps copying torch/include/ATen/native/mps/MPSGraphVenturaOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps copying torch/include/ATen/native/mps/OperationUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps copying torch/include/ATen/native/mps/TensorFactory.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps copying torch/include/ATen/native/mps/UnaryConstants.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/mps creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested copying torch/include/ATen/native/nested/NestedTensorBinaryOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested copying torch/include/ATen/native/nested/NestedTensorFactories.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested copying torch/include/ATen/native/nested/NestedTensorMath.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested copying torch/include/ATen/native/nested/NestedTensorTransformerFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested copying torch/include/ATen/native/nested/NestedTensorTransformerUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested copying torch/include/ATen/native/nested/NestedTensorUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/nested creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/AffineQuantizer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/AffineQuantizerBase.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/ConvUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/Copy.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/FakeQuantAffine.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/IndexKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized copying torch/include/ATen/native/quantized/PackedParams.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/BinaryOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/EmbeddingPackedParams.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/OnednnUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/QnnpackUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/QuantUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/QuantizedOps.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/RuyUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/XnnpackUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/conv_serialization.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/fbgemm_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/init_qnnpack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/qembeddingbag.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu copying torch/include/ATen/native/quantized/cpu/qembeddingbag_prepack.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/quantized/cpu creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/transformers copying torch/include/ATen/native/transformers/attention.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/transformers copying torch/include/ATen/native/transformers/sdp_utils_cpp.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/transformers creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/utils copying torch/include/ATen/native/utils/Factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/utils copying torch/include/ATen/native/utils/ParamUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/utils copying torch/include/ATen/native/utils/ParamsHash.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/native/utils creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/quantized copying torch/include/ATen/quantized/QTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/quantized copying torch/include/ATen/quantized/Quantizer.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/quantized creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu copying torch/include/ATen/xpu/CachingHostAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu copying torch/include/ATen/xpu/PinnedMemoryAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu copying torch/include/ATen/xpu/XPUContext.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu copying torch/include/ATen/xpu/XPUDevice.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu copying torch/include/ATen/xpu/XPUEvent.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu copying torch/include/ATen/xpu/XPUGeneratorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu/detail copying torch/include/ATen/xpu/detail/XPUHooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/xpu/detail creating build/lib.linux-x86_64-cpython-313/torch/include/caffe2 creating build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/read_adapter_interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/in_memory_adapter.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/crc_alt.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/istream_adapter.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/file_adapter.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/inline_container.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize copying torch/include/caffe2/serialize/versions.h -> build/lib.linux-x86_64-cpython-313/torch/include/caffe2/serialize creating build/lib.linux-x86_64-cpython-313/torch/include/c10 creating build/lib.linux-x86_64-cpython-313/torch/include/c10/macros copying torch/include/c10/macros/Export.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/macros copying torch/include/c10/macros/Macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/macros copying torch/include/c10/macros/cmake_macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/macros creating build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/thread_pool.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Layout.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/AutogradState.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/ScalarType.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SymBool.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SymInt.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/ScalarTypeToTypeMeta.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/TensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/MemoryFormat.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/RefcountedDeleter.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DefaultTensorOptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/PyHandleCache.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Event.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SymbolicShapeMeta.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Allocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/TensorOptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DefaultDtype.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DynamicCast.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/CopyBytes.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Device.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Backend.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/QEngine.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SymIntArrayRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SymFloat.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Scalar.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DeviceArray.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DispatchKey.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/alignment.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/OptionalRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/GeneratorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DeviceGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/InferenceMode.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/GradMode.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/UndefinedTensorImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SafePyObject.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Stream.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/WrapDimMinimal.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Storage.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/SymNodeImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DeviceType.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/ConstantSymNodeImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/Contiguity.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/StreamGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/DispatchKeySet.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/CompileTimeFunctionPointer.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/CPUAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/StorageImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core copying torch/include/c10/core/QScheme.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing copying torch/include/ATen/core/boxing/BoxedKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing copying torch/include/ATen/core/boxing/BoxedKernel_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing copying torch/include/ATen/core/boxing/KernelFunction.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing copying torch/include/ATen/core/boxing/KernelFunction_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing copying torch/include/ATen/core/boxing/OperatorKernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing/impl copying torch/include/ATen/core/boxing/impl/WrapFunctionIntoFunctor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing/impl copying torch/include/ATen/core/boxing/impl/WrapFunctionIntoRuntimeFunctor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing/impl copying torch/include/ATen/core/boxing/impl/boxing.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing/impl copying torch/include/ATen/core/boxing/impl/make_boxed_from_unboxed_functor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing/impl copying torch/include/ATen/core/boxing/impl/test_helpers.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/boxing/impl creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/CppSignature.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/DispatchKeyExtractor.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/Dispatcher.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/ObservedOperators.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/OperatorEntry.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/OperatorOptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch copying torch/include/ATen/core/dispatch/RegistrationHandleRAII.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/dispatch creating build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/op_registration copying torch/include/ATen/core/op_registration/adaption.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/op_registration copying torch/include/ATen/core/op_registration/infer_schema.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/op_registration copying torch/include/ATen/core/op_registration/op_allowlist.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/op_registration copying torch/include/ATen/core/op_registration/op_registration.h -> build/lib.linux-x86_64-cpython-313/torch/include/ATen/core/op_registration creating build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/COWDeleter.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/TorchDispatchModeTLS.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/COW.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/SizesAndStrides.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/FakeGuardImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/LocalDispatchKeySet.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/InlineStreamGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/VirtualGuardImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/DeviceGuardImplInterface.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/HermeticPyObjectTLS.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/PyObjectSlot.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/GPUTrace.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/alloc_cpu.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/InlineDeviceGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/PythonDispatcherTLS.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/PyInterpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl copying torch/include/c10/core/impl/InlineEvent.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/core/impl creating build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e5m2fnuz.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/llvmMathExtras.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Metaprogramming.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e4m3fn-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/BFloat16.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/AlignOf.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/TypeIndex.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/UniqueVoidPtr.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/SmallVector.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/BFloat16-math.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/floating_point_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ThreadLocalDebugInfo.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/complex_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/qint32.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/bits.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/order_preserving_flat_hash_map.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ConstexprCrc.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/sparse_bitset.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/copysign.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/LeftRight.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e5m2-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/thread_name.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e4m3fn.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/intrusive_ptr.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ExclusivelyOwned.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/AbortHandler.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/logging_is_not_google_glog.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Half-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Load.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/quint4x2.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/string_view.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/MathConstants.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Array.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/IdWrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ExclusivelyOwnedTensorTraits.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ssize.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/signal_handler.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/accumulate.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/env.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/generic_math.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/TypeList.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/logging_is_google_glog.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ParallelGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/OptionalArrayRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/qint8.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Unroll.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/win32-headers.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/hash.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/TypeCast.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/typeid.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/TypeTraits.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/int128.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/numa.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/complex_math.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/quint2x4.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ArrayRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Half.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/BFloat16-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/quint8.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e4m3fnuz-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/StringUtil.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/complex.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e5m2.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/string_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/strides.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/C++17.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/tempfile.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ApproximateClock.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Optional.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/static_tracepoint.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/CallOnce.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ThreadLocal.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Deprecated.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/strong_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Backtrace.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/flat_hash_map.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/DeadlockDetection.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Lazy.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Synchronized.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/irange.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/SmallBuffer.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Bitset.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e5m2fnuz-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/FbcodeMaps.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/DimVector.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/overloaded.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_e4m3fnuz.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/static_tracepoint_elfx86.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Type.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/FunctionRef.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/ScopeExit.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/safe_numerics.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Flags.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/MaybeOwned.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Exception.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Unicode.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Logging.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/python_stub.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/Float8_fnuz_cvt.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/bit_cast.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util copying torch/include/c10/util/TypeSafeSignMath.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/util creating build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAException.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDADeviceAssertion.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAStream.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/driver_api.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDACachingAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDADeviceAssertionHost.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAGraphsC10Utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAMacros.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAAlgorithm.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAAllocatorConfig.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAMiscFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda copying torch/include/c10/cuda/CUDAMathCompat.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda creating build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda/impl copying torch/include/c10/cuda/impl/CUDATest.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda/impl copying torch/include/c10/cuda/impl/CUDAGuardImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/cuda/impl creating build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPException.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPDeviceAssertion.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPStream.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/driver_api.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPCachingAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPDeviceAssertionHost.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPGraphsC10Utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPGuard.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPMacros.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPAlgorithm.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPAllocatorConfig.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPMiscFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip copying torch/include/c10/hip/HIPMathCompat.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip creating build/lib.linux-x86_64-cpython-313/torch/include/c10/hip/impl copying torch/include/c10/hip/impl/hip_cmake_macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip/impl copying torch/include/c10/hip/impl/HIPTest.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip/impl copying torch/include/c10/hip/impl/HIPGuardImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/hip/impl creating build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu copying torch/include/c10/xpu/XPUFunctions.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu copying torch/include/c10/xpu/XPUCachingAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu copying torch/include/c10/xpu/XPUException.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu copying torch/include/c10/xpu/XPUDeviceProp.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu copying torch/include/c10/xpu/XPUMacros.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu copying torch/include/c10/xpu/XPUStream.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu creating build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu/impl copying torch/include/c10/xpu/impl/XPUGuardImpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/c10/xpu/impl creating build/lib.linux-x86_64-cpython-313/torch/include/torch copying torch/include/torch/custom_class.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch copying torch/include/torch/custom_class_detail.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch copying torch/include/torch/library.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch copying torch/include/torch/script.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch copying torch/include/torch/extension.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Export.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Layout.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/python_headers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/MemoryFormat.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/DynamicTypes.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/DataLoader.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/StorageMethods.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/serialization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Event.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/THConcat.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Generator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Device.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Dtype.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/CudaIPCTypes.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/StorageSharing.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Stream.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Storage.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/THP.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Size.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/copy_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/TypeInfo.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/Module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/itt_wrapper.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/PyInterpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/python_dimname.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc copying torch/include/torch/csrc/QScheme.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/all.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/enum.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/nn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/python.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/xpu.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/sparse.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/optim.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/ordered_dict.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/imethod.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/jit.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/torch.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/fft.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/mps.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/arg.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/special.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/expanding_array.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/linalg.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/cuda.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/nested.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/serialize.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/data.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch copying torch/include/torch/csrc/api/include/torch/version.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/transforms.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/worker_exception.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/iterator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/samplers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/datasets.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/dataloader.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/dataloader_options.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data copying torch/include/torch/csrc/api/include/torch/data/example.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/dataloader copying torch/include/torch/csrc/api/include/torch/data/dataloader/stateful.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/dataloader copying torch/include/torch/csrc/api/include/torch/data/dataloader/stateless.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/dataloader copying torch/include/torch/csrc/api/include/torch/data/dataloader/base.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/dataloader creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/stateful.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/mnist.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/chunk.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/shared.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/map.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets copying torch/include/torch/csrc/api/include/torch/data/datasets/base.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/datasets creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/detail copying torch/include/torch/csrc/api/include/torch/data/detail/queue.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/detail copying torch/include/torch/csrc/api/include/torch/data/detail/data_shuttle.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/detail copying torch/include/torch/csrc/api/include/torch/data/detail/sequencers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/detail creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/distributed.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/stream.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/sequential.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/random.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/base.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/custom_batch_request.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers copying torch/include/torch/csrc/api/include/torch/data/samplers/serialize.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/samplers creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/transforms copying torch/include/torch/csrc/api/include/torch/data/transforms/lambda.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/transforms copying torch/include/torch/csrc/api/include/torch/data/transforms/collate.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/transforms copying torch/include/torch/csrc/api/include/torch/data/transforms/stack.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/transforms copying torch/include/torch/csrc/api/include/torch/data/transforms/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/transforms copying torch/include/torch/csrc/api/include/torch/data/transforms/base.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/data/transforms creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/detail copying torch/include/torch/csrc/api/include/torch/detail/TensorDataContainer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/detail copying torch/include/torch/csrc/api/include/torch/detail/static.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/detail creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/modules.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/cloneable.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/pimpl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/pimpl-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/functional.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn copying torch/include/torch/csrc/api/include/torch/nn/options.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/batchnorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/pooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/fold.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/normalization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/instancenorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/vision.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/upsampling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/pixelshuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/padding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/embedding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/conv.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional copying torch/include/torch/csrc/api/include/torch/nn/functional/distance.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/functional creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/transformercoder.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/batchnorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/pooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/transformer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/fold.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/normalization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/transformerlayer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/rnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/instancenorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/vision.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/upsampling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/pixelshuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/padding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/embedding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/conv.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/adaptive.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options copying torch/include/torch/csrc/api/include/torch/nn/options/distance.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/options creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/transformercoder.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/loss.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/batchnorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/common.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/pooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/transformer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/fold.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/normalization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/transformerlayer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/activation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/rnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/instancenorm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/upsampling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/pixelshuffle.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/padding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/embedding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/conv.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/adaptive.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules copying torch/include/torch/csrc/api/include/torch/nn/modules/distance.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/parameterdict.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/named_any.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/moduledict.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/parameterlist.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/any_value.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/functional.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/any_module_holder.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/any.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/modulelist.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container copying torch/include/torch/csrc/api/include/torch/nn/modules/container/sequential.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/modules/container creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/parallel copying torch/include/torch/csrc/api/include/torch/nn/parallel/data_parallel.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/parallel creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/utils copying torch/include/torch/csrc/api/include/torch/nn/utils/clip_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/utils copying torch/include/torch/csrc/api/include/torch/nn/utils/convert_parameters.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/utils copying torch/include/torch/csrc/api/include/torch/nn/utils/rnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/nn/utils creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/adamw.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/adam.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/lbfgs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/optimizer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/sgd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/rmsprop.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/serialize.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim copying torch/include/torch/csrc/api/include/torch/optim/adagrad.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim/schedulers copying torch/include/torch/csrc/api/include/torch/optim/schedulers/reduce_on_plateau_scheduler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim/schedulers copying torch/include/torch/csrc/api/include/torch/optim/schedulers/step_lr.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim/schedulers copying torch/include/torch/csrc/api/include/torch/optim/schedulers/lr_scheduler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/optim/schedulers creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/serialize copying torch/include/torch/csrc/api/include/torch/serialize/archive.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/serialize copying torch/include/torch/csrc/api/include/torch/serialize/input-archive.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/serialize copying torch/include/torch/csrc/api/include/torch/serialize/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/serialize copying torch/include/torch/csrc/api/include/torch/serialize/output-archive.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/api/include/torch/serialize creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/saved_variable_hooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_special_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_anomaly_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/autograd_not_implemented_fallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/saved_variable.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/jit_decomp_interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/profiler_legacy.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/variable_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_torch_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_variable_indexing.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_legacy_variable.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_saved_variable_hooks.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_hook.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/VariableTypeUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_nested_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/record_function_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_function.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_variable.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_cpp_function.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/custom_function.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/InferenceMode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/input_buffer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/engine.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_linalg_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/function.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/forward_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/graph_task.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/FunctionsManual.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/edge.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/variable.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/grad_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_fft_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_sparse_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_enum_tag.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/anomaly_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_nn_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_engine.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/profiler_python.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/function_hook.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/profiler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/python_autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/profiler_kineto.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/symbolic.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/cpp_hook.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd copying torch/include/torch/csrc/autograd/input_metadata.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions copying torch/include/torch/csrc/autograd/functions/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions copying torch/include/torch/csrc/autograd/functions/pybind.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions copying torch/include/torch/csrc/autograd/functions/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions copying torch/include/torch/csrc/autograd/functions/comm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions copying torch/include/torch/csrc/autograd/functions/basic_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions copying torch/include/torch/csrc/autograd/functions/accumulate_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/functions creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated copying torch/include/torch/csrc/autograd/generated/python_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated copying torch/include/torch/csrc/autograd/generated/python_return_types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated copying torch/include/torch/csrc/autograd/generated/VariableType.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated copying torch/include/torch/csrc/autograd/generated/Functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated copying torch/include/torch/csrc/autograd/generated/variable_factories.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated copying torch/include/torch/csrc/autograd/generated/ViewFuncs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/generated creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils copying torch/include/torch/csrc/autograd/utils/python_arg_parsing.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils copying torch/include/torch/csrc/autograd/utils/grad_layout_contract.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils copying torch/include/torch/csrc/autograd/utils/lambda_post_hook.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils copying torch/include/torch/csrc/autograd/utils/wrap_outputs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils copying torch/include/torch/csrc/autograd/utils/warnings.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils copying torch/include/torch/csrc/autograd/utils/error_messages.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/autograd/utils creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/Event.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/CUDAPluggableAllocator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/THCP.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/comm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/python_comm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/nccl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/Stream.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/device_set.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/memory_snapshot.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/python_nccl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda copying torch/include/torch/csrc/cuda/Module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/cuda creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/c10d.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/socket.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/TraceUtils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/logging.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/python_comm_hook.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/debug.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/exception.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/WinSockUtils.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/RankLocal.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/UnixSockUtils.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/UCCUtils.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/intra_node_comm.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupGloo.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/Utils.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/comm.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/Types.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/logger.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/GlooDeviceFactory.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/TCPStoreBackend.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroup.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/NCCLUtils.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/reducer_timer.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/Store.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ParamCommsUtils.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/PrefixStore.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/FakeProcessGroup.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/Backend.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/PyProcessGroup.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupCudaP2P.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/Work.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupRoundRobin.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/GroupRegistry.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/FileStore.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/HashStore.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/sequence_num.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/TCPStore.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/UCCTracing.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupWrapper.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupNCCL.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupUCC.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/reducer.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/ProcessGroupMPI.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d copying torch/include/torch/csrc/distributed/c10d/default_comm_hooks.hpp -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/c10d creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/rpc.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/rref_proto.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/python_call.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/message.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/tensorpipe_agent.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/agent_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/tensorpipe_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/rref_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/python_rpc_handler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/py_rref.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/rpc_command_base.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/python_remote_call.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/script_remote_call.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/script_call.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/request_callback_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/request_callback.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/rref_context.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/script_resp.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/torchscript_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/rpc_agent.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/request_callback_no_python.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/python_resp.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/unpickled_python_call.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/unpickled_python_remote_call.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc copying torch/include/torch/csrc/distributed/rpc/python_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/rpc creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/context copying torch/include/torch/csrc/distributed/autograd/context/context.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/context copying torch/include/torch/csrc/distributed/autograd/context/container.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/context creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/functions copying torch/include/torch/csrc/distributed/autograd/functions/sendrpc_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/functions copying torch/include/torch/csrc/distributed/autograd/functions/recvrpc_backward.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/functions creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_resp.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_req.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/rref_backward_resp.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/rpc_with_autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_req.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/cleanup_autograd_context_resp.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/rref_backward_req.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/autograd_metadata.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/rpc_with_profiling_resp.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages copying torch/include/torch/csrc/distributed/autograd/rpc_messages/propagate_gradients_req.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/distributed/autograd/rpc_messages creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/extra_state.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/eval_frame.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/python_compiled_autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/compiled_autograd.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/cpp_shim.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/cpython_defs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/debug_macros.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/guards.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo copying torch/include/torch/csrc/dynamo/cache_entry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/dynamo creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor copying torch/include/torch/csrc/inductor/inductor_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runner copying torch/include/torch/csrc/inductor/aoti_runner/model_container_runner.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runner copying torch/include/torch/csrc/inductor/aoti_runner/pybind.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runner copying torch/include/torch/csrc/inductor/aoti_runner/model_container_runner_cpu.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runner copying torch/include/torch/csrc/inductor/aoti_runner/model_container_runner_cuda.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runner creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/model_container.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/thread_local.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/arrayref_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/utils_cuda.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/scalar_to_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/model.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime copying torch/include/torch/csrc/inductor/aoti_runtime/device_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_runtime creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch copying torch/include/torch/csrc/inductor/aoti_torch/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch copying torch/include/torch/csrc/inductor/aoti_torch/tensor_converter.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch copying torch/include/torch/csrc/inductor/aoti_torch/proxy_executor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch copying torch/include/torch/csrc/inductor/aoti_torch/mkldnn_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch/c copying torch/include/torch/csrc/inductor/aoti_torch/c/shim.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch/c creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch/generated copying torch/include/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch/generated copying torch/include/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/inductor/aoti_torch/generated creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit copying torch/include/torch/csrc/jit/resource_guard.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit copying torch/include/torch/csrc/jit/jit_opt_limit.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit copying torch/include/torch/csrc/jit/jit_log.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_debug_handler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_exception.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_preprocess.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_resolver.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_debug_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends copying torch/include/torch/csrc/jit/backends/backend_detail.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/backends creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/remove_mutation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/loop_unrolling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/annotate_warns.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/metal_rewrite.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/guard_elimination.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/remove_inplace_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/inline_fork_wait.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_linear_folding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/fold_linear_bn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/add_if_then_else.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/constant_pooling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/xnnpack_rewrite.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/mkldnn_rewrite.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/autocast.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/concat_opt.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/specialize_autogradzero.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/device_type_analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/peephole_alias_sensitive.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/inliner.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/restore_mutation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/common_subexpression_elimination.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/erase_number_types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/peephole_dict_idioms.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/clear_profiling.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/remove_redundant_profiles.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/fuse_linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/create_autodiff_subgraphs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/tensorexpr_fuser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/value_refinement_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/inline_autodiff_subgraphs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/inline_forked_closures.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/remove_expands.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/constant_propagation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/lift_closures.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/bailout_graph.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/freeze_module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/liveness.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_ops_to_mkldnn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_conv_add_relu_fusion.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/fold_conv_bn.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/onnx.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/replacement_of_old_operators.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/vulkan_rewrite.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/remove_dropout.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/graph_fuser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_concat_linear.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/dead_code_elimination.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/batch_mm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/canonicalize.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/symbolic_shape_runtime_fusion.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/pass_manager.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/clear_undefinedness.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/onednn_graph_fuser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/peephole_list_idioms.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/peephole.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/symbolic_shape_cache.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_conv_folding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/decompose_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/refine_tuple_types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/lower_graph.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/lower_tuples.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/dtype_analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/hoist_conv_packed_params.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/fixup_trace_scope_blocks.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/remove_exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/normalize_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/eliminate_no_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/lower_grad_of.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/update_differentiable_graph_requires_grad.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/peephole_non_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/symbolic_shape_analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_linear_transpose.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/inplace_check.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/check_strict_fusion.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/create_functional_graphs.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/variadic_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/graph_rewrite_helper.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/canonicalize_graph_fuser_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/prepack_folding.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/subgraph_rewrite.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/insert_guards.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/frozen_graph_optimizations.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/mobile_optimizer_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/fuse_relu.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/integer_value_refinement.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/requires_grad_analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes copying torch/include/torch/csrc/jit/passes/shape_analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/quantization_patterns.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/dedup_module_uses.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/quantization_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/register_packed_params.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/fusion_passes.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/insert_observers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/helper.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/finalize.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization copying torch/include/torch/csrc/jit/passes/quantization/insert_quant_dequant.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/quantization creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/utils copying torch/include/torch/csrc/jit/passes/utils/check_alias_annotation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/utils copying torch/include/torch/csrc/jit/passes/utils/memory_dag.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/utils copying torch/include/torch/csrc/jit/passes/utils/subgraph_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/utils copying torch/include/torch/csrc/jit/passes/utils/optimization_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/utils copying torch/include/torch/csrc/jit/passes/utils/op_registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/passes/utils creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/calculate_necessary_args.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/argument_spec.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/serialized_shape_function_registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/script_profile.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/custom_operator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/interpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/vararg_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/exception_message.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/graph_executor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/simple_graph_executor_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/profiling_graph_executor_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/slice_indices_adjust.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/graph_iterator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/register_ops_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/shape_function_registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/print_handler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/decomposition_registry_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/variable_tensor_list.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/operator_options.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/graph_executor_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/profiling_record.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/logging.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/instruction.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/decomposition_registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/symbolic_script.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/jit_exception.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/operator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/jit_trace.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/symbolic_shape_registry_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/autodiff.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime copying torch/include/torch/csrc/jit/runtime/symbolic_shape_registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/runtime creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/attributes.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/subgraph_matcher.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/irparser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/constants.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/alias_analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/named_value.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/node_hashing.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/ir.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/graph_node_list.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/ir_views.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/scope.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/type_hashing.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir copying torch/include/torch/csrc/jit/ir/graph_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/ir creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/source_range.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/lexer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/strtod.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/parser_constants.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/function_schema_parser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/parse_string_literal.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/schema_type_parser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/error_report.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/tree.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/inline_loop_condition.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/exit_transforms.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/source_ref.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/mini_environment.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/convert_to_ssa.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/edit_distance.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/versioned_symbols.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/script_type_parser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/resolver.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/tree_views.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/sugared_value.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/name_mangler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/builtin_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/canonicalize_modified_loop.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/ir_emitter.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/tracer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/concrete_module_type.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/schema_matching.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend copying torch/include/torch/csrc/jit/frontend/parser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/frontend creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/api copying torch/include/torch/csrc/jit/api/function_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/api copying torch/include/torch/csrc/jit/api/module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/api copying torch/include/torch/csrc/jit/api/compilation_unit.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/api copying torch/include/torch/csrc/jit/api/object.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/api copying torch/include/torch/csrc/jit/api/method.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/api creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/callstack_debug_info_serialization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/source_range_serialization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/import.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/storage_context.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/import_source.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/pickle.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/import_read.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/flatbuffer_serializer_jit.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/onnx.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/pickler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/import_export_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/source_range_serialization_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/type_name_uniquer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/flatbuffer_serializer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/mobile_bytecode_generated.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/export.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/import_export_constants.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/export_bytecode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/python_print.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/unpickler.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization copying torch/include/torch/csrc/jit/serialization/import_export_helpers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/serialization creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_custom_class.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/pybind_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/update_graph_executor_opt.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/utf8_decoding_ignore.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/pybind.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_arg_flatten.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_list.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/module_python.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_dict.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/script_init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_tracer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_sugared_value.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_ivalue.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_tree_views.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python copying torch/include/torch/csrc/jit/python/python_ir.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/python creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/file_format.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/parse_operators.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/interpreter.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/import.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/flatbuffer_loader.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/register_ops_common_utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/code.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/promoted_prim_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/type_parser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/profiler_edge.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/import_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/upgrader_mobile.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/debug_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/quantization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/frame.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/parse_bytecode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/function.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/observer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/method.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/import_export_common.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile copying torch/include/torch/csrc/jit/mobile/prim_ops_registery.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/mobile creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/testing copying torch/include/torch/csrc/jit/testing/hooks_for_testing.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/testing copying torch/include/torch/csrc/jit/testing/file_check.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/testing creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/intrinsic_symbols.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/block_codegen.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/mem_dependency_checker.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/expr.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/hash_provider.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/kernel.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/cpp_codegen.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/half_support.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/var_substitutor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir_cloner.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/lowerings.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/exceptions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir_visitor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/fwd_decls.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/external_functions_registry.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/reduction.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/registerizer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/eval.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/external_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/cuda_random.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/loopnest_randomization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/loopnest.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/stmt.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/llvm_jit.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/llvm_codegen.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/graph_opt.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir_verifier.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/bounds_inference.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/cuda_codegen.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/tensorexpr_init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/codegen.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir_printer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir_simplifier.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/cpp_intrinsics.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/external_functions_core.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/analysis.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/unique_name_manager.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/ir_mutator.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr copying torch/include/torch/csrc/jit/tensorexpr/bounds_overlap.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/pointwise.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/norm.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/operators.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/conv2d.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/matmul.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/reduction.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/quantization.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/softmax.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators copying torch/include/torch/csrc/jit/tensorexpr/operators/misc.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/tensorexpr/operators creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/codegen creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/codegen/cuda copying torch/include/torch/csrc/jit/codegen/cuda/interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/jit/codegen/cuda creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/onnx copying torch/include/torch/csrc/onnx/back_compat.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/onnx copying torch/include/torch/csrc/onnx/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/onnx copying torch/include/torch/csrc/onnx/onnx.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/onnx creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/collection.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/api.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/kineto_shim.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/data_flow.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/events.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/containers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/perf.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/perf-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler copying torch/include/torch/csrc/profiler/combined_traceback.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/orchestration copying torch/include/torch/csrc/profiler/orchestration/vulkan.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/orchestration copying torch/include/torch/csrc/profiler/orchestration/observer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/orchestration copying torch/include/torch/csrc/profiler/orchestration/python_tracer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/orchestration creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/standalone copying torch/include/torch/csrc/profiler/standalone/nvtx_observer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/standalone copying torch/include/torch/csrc/profiler/standalone/privateuse1_observer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/standalone copying torch/include/torch/csrc/profiler/standalone/execution_trace_observer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/standalone copying torch/include/torch/csrc/profiler/standalone/itt_observer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/standalone creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/stubs copying torch/include/torch/csrc/profiler/stubs/base.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/stubs creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/dwarf_symbolize_enums.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/range_table.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/dwarf_enums.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/lexer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/unwind_error.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/fast_symbolizer.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/mem_file.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/fde.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/action.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/unwinder.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/sections.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/eh_frame_hdr.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/debug_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/communicate.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/line_number_program.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind copying torch/include/torch/csrc/profiler/unwind/unwind.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/unwind creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/python copying torch/include/torch/csrc/profiler/python/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/python copying torch/include/torch/csrc/profiler/python/pybind.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/python copying torch/include/torch/csrc/profiler/python/combined_traceback.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/profiler/python creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_memoryformats.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/schema_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/disable_torch_function.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/throughput_benchmark.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/pythoncapi_compat.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/throughput_benchmark-inl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_compat.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/cpp_stacktraces.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/pybind.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_qschemes.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_list.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/cuda_enabled.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_torch_function_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_numbers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_scalars.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_arg_parser.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/out_types.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/pycfunction_helpers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_new.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_tuples.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/verbose.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/six.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_layouts.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/pyobject_preservation.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/object_ptr.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_symnode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/structseq.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/numpy_stub.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_dtypes.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_apply.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_flatten.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/torch_dispatch_mode.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/nested.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/tensor_numpy.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/byte_order.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_raii.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/device_lazy_init.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/variadic.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/invalid_arguments.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_strings.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_stub.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils copying torch/include/torch/csrc/utils/python_dispatch.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/utils creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/tensor copying torch/include/torch/csrc/tensor/python_tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/tensor creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/backend copying torch/include/torch/csrc/lazy/backend/backend_device.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/backend copying torch/include/torch/csrc/lazy/backend/backend_interface.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/backend copying torch/include/torch/csrc/lazy/backend/backend_data.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/backend copying torch/include/torch/csrc/lazy/backend/lowering_context.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/backend creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/thread_pool.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/helpers.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/multi_wait.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/ir_builder.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/shape_inference.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/permutation_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/trie.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/debug_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/tensor_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/ir_dump_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/ir_metadata.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/tensor_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/hash.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/metrics.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/unique.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/ir.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/cache.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/shape.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/ir_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/dynamic_ir.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/config.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core copying torch/include/torch/csrc/lazy/core/lazy_graph_executor.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core/internal_ops copying torch/include/torch/csrc/lazy/core/internal_ops/ltc_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core/internal_ops creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core/ops copying torch/include/torch/csrc/lazy/core/ops/utils.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core/ops copying torch/include/torch/csrc/lazy/core/ops/arithmetic_ir_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/core/ops creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/python copying torch/include/torch/csrc/lazy/python/python_util.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/python creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/tensor_aten_ops.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ir_builder.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ts_node_lowering.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ts_lowering_context.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ts_node.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ts_autograd_functions.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ts_backend_impl.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/dynamic_ir.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/config.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend copying torch/include/torch/csrc/lazy/ts_backend/ts_eager_fallback.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/lazy/ts_backend creating build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/xpu copying torch/include/torch/csrc/xpu/Event.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/xpu copying torch/include/torch/csrc/xpu/Stream.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/xpu copying torch/include/torch/csrc/xpu/Module.h -> build/lib.linux-x86_64-cpython-313/torch/include/torch/csrc/xpu creating build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/attr.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/buffer_info.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/cast.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/chrono.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/common.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/complex.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/eigen.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/embed.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/eval.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/functional.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/gil.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/iostream.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/numpy.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/operators.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/options.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/pybind11.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/pytypes.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/stl.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/stl_bind.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 copying torch/include/pybind11/type_caster_pyobject_ptr.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11 creating build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/class.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/common.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/descr.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/init.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/internals.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/type_caster_base.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail copying torch/include/pybind11/detail/typeid.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/detail creating build/lib.linux-x86_64-cpython-313/torch/include/pybind11/eigen copying torch/include/pybind11/eigen/common.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/eigen copying torch/include/pybind11/eigen/matrix.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/eigen copying torch/include/pybind11/eigen/tensor.h -> build/lib.linux-x86_64-cpython-313/torch/include/pybind11/eigen creating build/lib.linux-x86_64-cpython-313/torch/include/THH copying torch/include/THH/THHAtomics.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/THH copying torch/include/THH/THHDeviceUtils.cuh -> build/lib.linux-x86_64-cpython-313/torch/include/THH copying torch/_inductor/codegen/cpp_prefix.h -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen creating build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/aoti_runtime copying torch/_inductor/codegen/aoti_runtime/interface.cpp -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/aoti_runtime copying torch/_inductor/codegen/aoti_runtime/implementation.cpp -> build/lib.linux-x86_64-cpython-313/torch/_inductor/codegen/aoti_runtime copying torch/_export/serde/schema.yaml -> build/lib.linux-x86_64-cpython-313/torch/_export/serde creating build/lib.linux-x86_64-cpython-313/torch/share creating build/lib.linux-x86_64-cpython-313/torch/share/cmake creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/ATen copying torch/share/cmake/ATen/ATenConfig.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/ATen creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 copying torch/share/cmake/Caffe2/Caffe2Config.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 copying torch/share/cmake/Caffe2/FindCUDAToolkit.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 copying torch/share/cmake/Caffe2/FindCUSPARSELT.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 copying torch/share/cmake/Caffe2/FindSYCLToolkit.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 copying torch/share/cmake/Caffe2/Caffe2Targets.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 copying torch/share/cmake/Caffe2/Caffe2Targets-relwithdebinfo.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2 creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/cuda.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/xpu.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/glog.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/gflags.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/mkl.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/mkldnn.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/protobuf.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/utils.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public copying torch/share/cmake/Caffe2/public/LoadHIP.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/public creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix copying torch/share/cmake/Caffe2/Modules_CUDA_fix/FindCUDNN.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix copying torch/share/cmake/Caffe2/Modules_CUDA_fix/FindCUDA.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindPackageHandleStandardArgs.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindPackageMessage.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/CMakeInitializeConfigs.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA/select_compute_arch.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA/make2cmake.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA/parse_cubin.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA copying torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA/run_nvcc.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Caffe2/Modules_CUDA_fix/upstream/FindCUDA creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Tensorpipe copying torch/share/cmake/Tensorpipe/TensorpipeTargets.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Tensorpipe copying torch/share/cmake/Tensorpipe/TensorpipeTargets-relwithdebinfo.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Tensorpipe creating build/lib.linux-x86_64-cpython-313/torch/share/cmake/Torch copying torch/share/cmake/Torch/TorchConfigVersion.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Torch copying torch/share/cmake/Torch/TorchConfig.cmake -> build/lib.linux-x86_64-cpython-313/torch/share/cmake/Torch copying torch/utils/benchmark/utils/timeit_template.cpp -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils copying torch/utils/benchmark/utils/valgrind_wrapper/compat_bindings.cpp -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper copying torch/utils/benchmark/utils/valgrind_wrapper/timer_callgrind_template.cpp -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper copying torch/utils/benchmark/utils/valgrind_wrapper/callgrind.h -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper copying torch/utils/benchmark/utils/valgrind_wrapper/valgrind.h -> build/lib.linux-x86_64-cpython-313/torch/utils/benchmark/utils/valgrind_wrapper copying torch/utils/model_dump/skeleton.html -> build/lib.linux-x86_64-cpython-313/torch/utils/model_dump copying torch/utils/model_dump/code.js -> build/lib.linux-x86_64-cpython-313/torch/utils/model_dump copying torch/utils/model_dump/preact.mjs -> build/lib.linux-x86_64-cpython-313/torch/utils/model_dump copying torch/utils/model_dump/htm.mjs -> build/lib.linux-x86_64-cpython-313/torch/utils/model_dump creating build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libc10_hip.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libc10.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libcaffe2_nvrtc.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libshm.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libtorch_python.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libtorch_cpu.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libtorch_hip.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libtorch.so -> build/lib.linux-x86_64-cpython-313/torch/lib copying torch/lib/libtorch_global_deps.so -> build/lib.linux-x86_64-cpython-313/torch/lib creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe copying torch/include/tensorpipe/tensorpipe.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe copying torch/include/tensorpipe/config.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel copying torch/include/tensorpipe/channel/context.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel copying torch/include/tensorpipe/channel/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/basic copying torch/include/tensorpipe/channel/basic/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/basic creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/cma copying torch/include/tensorpipe/channel/cma/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/cma creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/mpt copying torch/include/tensorpipe/channel/mpt/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/mpt creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/xth copying torch/include/tensorpipe/channel/xth/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/channel/xth creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/common copying torch/include/tensorpipe/common/buffer.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/common copying torch/include/tensorpipe/common/cpu_buffer.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/common copying torch/include/tensorpipe/common/device.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/common copying torch/include/tensorpipe/common/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/common copying torch/include/tensorpipe/common/optional.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/common creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/core copying torch/include/tensorpipe/core/context.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/core copying torch/include/tensorpipe/core/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/core copying torch/include/tensorpipe/core/listener.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/core copying torch/include/tensorpipe/core/message.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/core copying torch/include/tensorpipe/core/pipe.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/core creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport copying torch/include/tensorpipe/transport/context.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport copying torch/include/tensorpipe/transport/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/ibv copying torch/include/tensorpipe/transport/ibv/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/ibv copying torch/include/tensorpipe/transport/ibv/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/ibv copying torch/include/tensorpipe/transport/ibv/utility.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/ibv creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/shm copying torch/include/tensorpipe/transport/shm/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/shm creating build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/uv copying torch/include/tensorpipe/transport/uv/error.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/uv copying torch/include/tensorpipe/transport/uv/factory.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/uv copying torch/include/tensorpipe/transport/uv/utility.h -> build/lib.linux-x86_64-cpython-313/torch/include/tensorpipe/transport/uv copying torch/jit/_script.pyi -> build/lib.linux-x86_64-cpython-313/torch/jit copying torch/utils/_config_typing.pyi -> build/lib.linux-x86_64-cpython-313/torch/utils copying torch/nn/functional.pyi -> build/lib.linux-x86_64-cpython-313/torch/nn copying torch/optim/_multi_tensor/__init__.pyi -> build/lib.linux-x86_64-cpython-313/torch/optim/_multi_tensor copying torch/utils/data/datapipes/datapipe.pyi -> build/lib.linux-x86_64-cpython-313/torch/utils/data/datapipes copying torch/distributed/optim/zero_redundancy_optimizer.pyi -> build/lib.linux-x86_64-cpython-313/torch/distributed/optim copying torch/nn/utils/rnn.pyi -> build/lib.linux-x86_64-cpython-313/torch/nn/utils running build_ext -- Building with NumPy bindings -- Not using cuDNN -- Not using CUDA -- Not using XPU -- Not using MKLDNN -- Not using NCCL -- Building with distributed package: -- USE_TENSORPIPE=True -- USE_GLOO=False -- USE_MPI=False -- Building Executorch -- Not using ITT Copying functorch._C from functorch/functorch.so to /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib.linux-x86_64-cpython-313/functorch/_C.cpython-313-x86_64-linux-gnu.so copying functorch/functorch.so -> /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/lib.linux-x86_64-cpython-313/functorch/_C.cpython-313-x86_64-linux-gnu.so building 'torch._C' extension creating build/temp.linux-x86_64-cpython-313 creating build/temp.linux-x86_64-cpython-313/torch creating build/temp.linux-x86_64-cpython-313/torch/csrc gcc -fno-strict-overflow -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -fcf-protection -fexceptions -fcf-protection -fexceptions -fcf-protection -fexceptions -O3 -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -fPIC -I/usr/include/python3.13 -c torch/csrc/stub.c -o build/temp.linux-x86_64-cpython-313/torch/csrc/stub.o -Wall -Wextra -Wno-strict-overflow -Wno-unused-parameter -Wno-missing-field-initializers -Wno-unknown-pragmas -fno-strict-aliasing -g gcc -shared -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer build/temp.linux-x86_64-cpython-313/torch/csrc/stub.o -L/builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/lib -L/usr/lib64 -ltorch_python -o build/lib.linux-x86_64-cpython-313/torch/_C.cpython-313-x86_64-linux-gnu.so -g -Wl,-rpath,$ORIGIN/lib ------------------------------------------------------------------------- | | | It is no longer necessary to use the 'build' or 'rebuild' targets | | | | To install: | | $ python setup.py install | | To develop locally: | | $ python setup.py develop | | To force cmake to re-generate native build files (off by default): | | $ python setup.py develop --cmake | | | ------------------------------------------------------------------------- + mv build build-default + module purge + local _mlredir=0 + '[' -n '' ']' + case " $@ " in + '[' 0 -eq 0 ']' + _module_raw purge ++ /usr/bin/tclsh /usr/share/Modules/libexec/modulecmd.tcl bash purge + eval 'unset __MODULES_LMCONFLICT; unset ROCM_BIN; unset _LMFILES_; unset LOADEDMODULES; unset ROCM_GPUS; unset ROCM_LIB; test 0;' ++ unset __MODULES_LMCONFLICT ++ unset ROCM_BIN ++ unset _LMFILES_ ++ unset LOADEDMODULES ++ unset ROCM_GPUS ++ unset ROCM_LIB ++ test 0 + _mlstatus=0 + return 0 + for gpu in gfx9 + module load rocm/gfx9 + local _mlredir=0 + '[' -n '' ']' + case " $@ " in + '[' 0 -eq 0 ']' + _module_raw load rocm/gfx9 ++ /usr/bin/tclsh /usr/share/Modules/libexec/modulecmd.tcl bash load rocm/gfx9 + eval 'LD_LIBRARY_PATH=/usr/lib64/rocm/gfx9/lib; export LD_LIBRARY_PATH; __MODULES_LMCONFLICT=rocm/gfx9\&rocm; export __MODULES_LMCONFLICT; PATH=/usr/lib64/rocm/gfx9/bin:/usr/share/Modules/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/sbin; export PATH; ROCM_BIN=/usr/lib64/rocm/gfx9/bin; export ROCM_BIN; _LMFILES_=/usr/share/modulefiles/rocm/gfx9; export _LMFILES_; LOADEDMODULES=rocm/gfx9; export LOADEDMODULES; PYTHONPATH=/usr/lib64/rocm/gfx9/lib64/python3.13/site-packages/; export PYTHONPATH; ROCM_GPUS=gfx900\;gfx906:xnack-\;gfx908:xnack-\;gfx90a:xnack+\;gfx90a:xnack-\;gfx942; export ROCM_GPUS; ROCM_LIB=/usr/lib64/rocm/gfx9/lib; export ROCM_LIB; test 0;' ++ LD_LIBRARY_PATH=/usr/lib64/rocm/gfx9/lib ++ export LD_LIBRARY_PATH ++ __MODULES_LMCONFLICT='rocm/gfx9&rocm' ++ export __MODULES_LMCONFLICT ++ PATH=/usr/lib64/rocm/gfx9/bin:/usr/share/Modules/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/sbin ++ export PATH ++ ROCM_BIN=/usr/lib64/rocm/gfx9/bin ++ export ROCM_BIN ++ _LMFILES_=/usr/share/modulefiles/rocm/gfx9 ++ export _LMFILES_ ++ LOADEDMODULES=rocm/gfx9 ++ export LOADEDMODULES ++ PYTHONPATH=/usr/lib64/rocm/gfx9/lib64/python3.13/site-packages/ ++ export PYTHONPATH ++ ROCM_GPUS='gfx900;gfx906:xnack-;gfx908:xnack-;gfx90a:xnack+;gfx90a:xnack-;gfx942' ++ export ROCM_GPUS ++ ROCM_LIB=/usr/lib64/rocm/gfx9/lib ++ export ROCM_LIB ++ test 0 + _mlstatus=0 + return 0 + export 'PYTORCH_ROCM_ARCH=gfx900;gfx906:xnack-;gfx908:xnack-;gfx90a:xnack+;gfx90a:xnack-;gfx942' + PYTORCH_ROCM_ARCH='gfx900;gfx906:xnack-;gfx908:xnack-;gfx90a:xnack+;gfx90a:xnack-;gfx942' + CFLAGS='-O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer ' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes ' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -sP' -- The CXX compiler identification is GNU 14.2.1 -- The C compiler identification is GNU 14.2.1 -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/g++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/gcc - skipped -- Detecting C compile features -- Detecting C compile features - done -- /usr/bin/g++ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/abi-check.cpp -o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/abi-check -- Determined _GLIBCXX_USE_CXX11_ABI=1 -- Not forcing any particular BLAS to be found -- Could not find ccache. Consider installing ccache to speed up compilation. -- Performing Test C_HAS_AVX_1 -- Performing Test C_HAS_AVX_1 - Failed -- Performing Test C_HAS_AVX_2 -- Performing Test C_HAS_AVX_2 - Success -- Performing Test C_HAS_AVX2_1 -- Performing Test C_HAS_AVX2_1 - Failed -- Performing Test C_HAS_AVX2_2 -- Performing Test C_HAS_AVX2_2 - Success -- Performing Test C_HAS_AVX512_1 -- Performing Test C_HAS_AVX512_1 - Failed -- Performing Test C_HAS_AVX512_2 -- Performing Test C_HAS_AVX512_2 - Success -- Performing Test CXX_HAS_AVX_1 -- Performing Test CXX_HAS_AVX_1 - Failed -- Performing Test CXX_HAS_AVX_2 -- Performing Test CXX_HAS_AVX_2 - Success -- Performing Test CXX_HAS_AVX2_1 -- Performing Test CXX_HAS_AVX2_1 - Failed -- Performing Test CXX_HAS_AVX2_2 -- Performing Test CXX_HAS_AVX2_2 - Success -- Performing Test CXX_HAS_AVX512_1 -- Performing Test CXX_HAS_AVX512_1 - Failed -- Performing Test CXX_HAS_AVX512_2 -- Performing Test CXX_HAS_AVX512_2 - Success -- Current compiler supports avx2 extension. Will build perfkernels. -- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX512_EXTENSIONS -- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX512_EXTENSIONS - Success -- Current compiler supports avx512f extension. Will build fbgemm. -- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY -- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY - Success -- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY -- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY - Success -- Performing Test COMPILER_SUPPORTS_RDYNAMIC -- Performing Test COMPILER_SUPPORTS_RDYNAMIC - Success -- Caffe2: Found protobuf with new-style protobuf targets. -- Caffe2 protobuf include directory: /usr/include -- Found Threads: TRUE -- Trying to find preferred BLAS backend of choice: MKL -- MKL_THREADING = OMP -- Looking for sys/types.h -- Looking for sys/types.h - found -- Looking for stdint.h -- Looking for stdint.h - found -- Looking for stddef.h -- Looking for stddef.h - found -- Check size of void* -- Check size of void* - done -- MKL_THREADING = OMP CMake Warning at cmake/Dependencies.cmake:201 (message): MKL could not be found. Defaulting to Eigen Call Stack (most recent call first): CMakeLists.txt:854 (include) CMake Warning at cmake/Dependencies.cmake:243 (message): Preferred BLAS (MKL) cannot be found, now searching for a general BLAS library Call Stack (most recent call first): CMakeLists.txt:854 (include) -- MKL_THREADING = OMP -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl - guide - pthread - m] -- Library mkl: not found -- MKL library not found -- Checking for [blis] -- Library blis: BLAS_blis_LIBRARY-NOTFOUND -- Checking for [Accelerate] -- Library Accelerate: BLAS_Accelerate_LIBRARY-NOTFOUND -- Checking for [vecLib] -- Library vecLib: BLAS_vecLib_LIBRARY-NOTFOUND -- Checking for [flexiblas] -- Library flexiblas: /usr/lib64/libflexiblas.so -- Looking for sgemm_ -- Looking for sgemm_ - found -- Performing Test BLAS_F2C_DOUBLE_WORKS -- Performing Test BLAS_F2C_DOUBLE_WORKS - Failed -- Performing Test BLAS_F2C_FLOAT_WORKS -- Performing Test BLAS_F2C_FLOAT_WORKS - Success -- Performing Test BLAS_USE_CBLAS_DOT -- Performing Test BLAS_USE_CBLAS_DOT - Success -- Looking for sbgemm_ -- Looking for sbgemm_ - not found -- Found a library with BLAS API (flexi). Full path: (/usr/lib64/libflexiblas.so) -- Using pocketfft in directory: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/pocketfft/ -- Found pthreadpool: /usr/lib64/libpthreadpool.so Found cpuinfo: /usr/lib64/libcpuinfo.so -- Found XNNPACK: /usr/lib64/libXNNPACK.so CMake Warning at cmake/Dependencies.cmake:760 (message): Turning USE_FAKELOWP off as it depends on USE_FBGEMM. Call Stack (most recent call first): CMakeLists.txt:854 (include) -- Found Numa: /usr/include -- Found Numa (include: /usr/include, library: /usr/lib64/libnuma.so) -- Found system Eigen at /usr/include/eigen3 -- Found Python: /usr/bin/python3 (found version "3.13.1") found components: Interpreter Development.Module NumPy -- Using third_party/pybind11. -- pybind11 include dirs: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/cmake/../third_party/pybind11/include -- Could NOT find OpenTelemetryApi (missing: OpenTelemetryApi_INCLUDE_DIRS) -- Using third_party/opentelemetry-cpp. -- opentelemetry api include dirs: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/cmake/../third_party/opentelemetry-cpp/api/include -- Checking for module 'mpi-c' -- Package 'mpi-c' not found -- Could NOT find MPI_C (missing: MPI_C_LIB_NAMES MPI_C_HEADER_DIR MPI_C_WORKS) -- Checking for module 'mpi-cxx' -- Package 'mpi-cxx' not found -- Could NOT find MPI_CXX (missing: MPI_CXX_LIB_NAMES MPI_CXX_HEADER_DIR MPI_CXX_WORKS) -- Could NOT find MPI (missing: MPI_C_FOUND MPI_CXX_FOUND) CMake Warning at cmake/Dependencies.cmake:945 (message): Not compiling with MPI. Suppress this warning with -DUSE_MPI=OFF Call Stack (most recent call first): CMakeLists.txt:854 (include) -- MKL_THREADING = OMP -- Check OMP with lib /usr/lib64/libomp.so and flags -fopenmp -v -- MKL_THREADING = OMP -- Check OMP with lib /usr/lib64/libomp.so and flags -fopenmp -v -- Found OpenMP_C: -fopenmp (found version "4.5") -- Found OpenMP_CXX: -fopenmp (found version "4.5") -- Found OpenMP: TRUE (found version "4.5") -- Adding OpenMP CXX_FLAGS: -fopenmp -- Will link against OpenMP libraries: /usr/lib64/libomp.so Building PyTorch for GPU arch: gfx900;gfx906:xnack-;gfx908:xnack-;gfx90a:xnack+;gfx90a:xnack-;gfx942 -- Found HIP: /usr (found version "6.2.41134-0") HIP VERSION: 6.2.41134-0 -- Caffe2: Header version is: 6.2.0 ***** ROCm version from rocm_version.h **** ROCM_VERSION_DEV: 6.2.0 ROCM_VERSION_DEV_MAJOR: 6 ROCM_VERSION_DEV_MINOR: 2 ROCM_VERSION_DEV_PATCH: 0 ROCM_VERSION_DEV_INT: 60200 HIP_VERSION_MAJOR: 6 HIP_VERSION_MINOR: 2 TORCH_HIP_VERSION: 602 ***** Library versions from dpkg ***** ***** Library versions from cmake find_package ***** hip VERSION: 6.2.41134 hsa-runtime64 VERSION: 1.14.0 amd_comgr VERSION: 2.8.0 rocrand VERSION: 3.1.0 hiprand VERSION: 2.11.0 rocblas VERSION: 4.2.1 hipblas VERSION: 2.2.0 hipblaslt VERSION: 0.8.0 miopen VERSION: 3.2.0 hipfft VERSION: 1.0.15 hipsparse VERSION: 3.1.1 rccl VERSION: 2.20.5 rocprim VERSION: 3.2.0 hipcub VERSION: 3.2.0 rocthrust VERSION: 3.1.0 hipsolver VERSION: 2.2.0 HIP is using new type enums INFOCompiling with HIP for AMD. TORCH_HIP_VERSION=602 is added as a compiler defines -- Disabling Kernel Assert for ROCm CMake Deprecation Warning at third_party/tensorpipe/third_party/libuv/CMakeLists.txt:1 (cmake_minimum_required): Compatibility with CMake < 3.5 will be removed from a future version of CMake. Update the VERSION argument value or use a ... suffix to tell CMake that the project does not need compatibility with older versions. -- Performing Test UV_LINT_W4 -- Performing Test UV_LINT_W4 - Failed -- Performing Test UV_LINT_NO_UNUSED_PARAMETER_MSVC -- Performing Test UV_LINT_NO_UNUSED_PARAMETER_MSVC - Failed -- Performing Test UV_LINT_NO_CONDITIONAL_CONSTANT_MSVC -- Performing Test UV_LINT_NO_CONDITIONAL_CONSTANT_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_EMPTY_TU_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_EMPTY_TU_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_FILE_SCOPE_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_FILE_SCOPE_MSVC - Failed -- Performing Test UV_LINT_NO_NONSTANDARD_NONSTATIC_DLIMPORT_MSVC -- Performing Test UV_LINT_NO_NONSTANDARD_NONSTATIC_DLIMPORT_MSVC - Failed -- Performing Test UV_LINT_NO_HIDES_LOCAL -- Performing Test UV_LINT_NO_HIDES_LOCAL - Failed -- Performing Test UV_LINT_NO_HIDES_PARAM -- Performing Test UV_LINT_NO_HIDES_PARAM - Failed -- Performing Test UV_LINT_NO_HIDES_GLOBAL -- Performing Test UV_LINT_NO_HIDES_GLOBAL - Failed -- Performing Test UV_LINT_NO_CONDITIONAL_ASSIGNMENT_MSVC -- Performing Test UV_LINT_NO_CONDITIONAL_ASSIGNMENT_MSVC - Failed -- Performing Test UV_LINT_NO_UNSAFE_MSVC -- Performing Test UV_LINT_NO_UNSAFE_MSVC - Failed -- Performing Test UV_LINT_WALL -- Performing Test UV_LINT_WALL - Success -- Performing Test UV_LINT_NO_UNUSED_PARAMETER -- Performing Test UV_LINT_NO_UNUSED_PARAMETER - Success -- Performing Test UV_LINT_STRICT_PROTOTYPES -- Performing Test UV_LINT_STRICT_PROTOTYPES - Success -- Performing Test UV_LINT_EXTRA -- Performing Test UV_LINT_EXTRA - Success -- Performing Test UV_LINT_UTF8_MSVC -- Performing Test UV_LINT_UTF8_MSVC - Failed -- Performing Test UV_F_STRICT_ALIASING -- Performing Test UV_F_STRICT_ALIASING - Success -- summary of build options: Install prefix: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch Target system: Linux Compiler: C compiler: /usr/bin/gcc CFLAGS: -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -- Found uv: 1.38.1 (found version "1.38.1") CMake Warning at cmake/Dependencies.cmake:1168 (message): TensorPipe doesn't yet support ROCm Call Stack (most recent call first): CMakeLists.txt:854 (include) -- Found onnx: /usr/lib64/libonnx.so /usr/lib64/libonnx_proto.so -- Found CUDA with FP16 support, compiling with torch.cuda.HalfTensor -- Removing -DNDEBUG from compile flags -- Checking prototype magma_get_sgeqrf_nb for MAGMA_V2 -- Checking prototype magma_get_sgeqrf_nb for MAGMA_V2 - False -- Compiling with MAGMA support -- MAGMA INCLUDE DIRECTORIES: /usr/include -- MAGMA LIBRARIES: /usr/lib64/libmagma.so -- MAGMA V2 check: 0 CMake Warning (dev) at cmake/Modules/FindARM.cmake:5 (EXEC_PROGRAM): Policy CMP0153 is not set: The exec_program command should not be called. Run "cmake --help-policy CMP0153" for policy details. Use the cmake_policy command to set the policy and suppress this warning. Use execute_process() instead. Call Stack (most recent call first): cmake/Dependencies.cmake:1416 (find_package) CMakeLists.txt:854 (include) This warning is for project developers. Use -Wno-dev to suppress it. -- Could not find hardware support for NEON on this machine. -- No OMAP3 processor on this machine. -- No OMAP4 processor on this machine. -- MKL_THREADING = OMP -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_intel - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf_lp64 - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_gnu_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_gf - mkl_intel_thread - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_sequential - mkl_core - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_sequential - mkl_core - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - gomp - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - gomp - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - iomp5 - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl_intel_lp64 - mkl_core - pthread - m - dl] -- Library mkl_intel_lp64: not found -- Checking for [mkl_intel - mkl_core - pthread - m - dl] -- Library mkl_intel: not found -- Checking for [mkl_gf_lp64 - mkl_core - pthread - m - dl] -- Library mkl_gf_lp64: not found -- Checking for [mkl_gf - mkl_core - pthread - m - dl] -- Library mkl_gf: not found -- Checking for [mkl - guide - pthread - m] -- Library mkl: not found -- MKL library not found -- Checking for [blis] -- Library blis: BLAS_blis_LIBRARY-NOTFOUND -- Checking for [Accelerate] -- Library Accelerate: BLAS_Accelerate_LIBRARY-NOTFOUND -- Checking for [vecLib] -- Library vecLib: BLAS_vecLib_LIBRARY-NOTFOUND -- Checking for [flexiblas] -- Library flexiblas: /usr/lib64/libflexiblas.so -- Found a library with BLAS API (flexi). Full path: (/usr/lib64/libflexiblas.so) -- Looking for cheev_ -- Looking for cheev_ - found -- Found a library with LAPACK API (flexi). disabling CUDA because NOT USE_CUDA is set disabling MKLDNN because USE_MKLDNN is not set -- Looking for clock_gettime in rt -- Looking for clock_gettime in rt - found -- Looking for mmap -- Looking for mmap - found -- Looking for shm_open -- Looking for shm_open - found -- Looking for shm_unlink -- Looking for shm_unlink - found -- Looking for malloc_usable_size -- Looking for malloc_usable_size - found -- -- Performing Test COMPILE_OUT_ZVECTOR -- Performing Test COMPILE_OUT_ZVECTOR - Failed -- ZVECTOR flags were NOT set. -- -- GCC 14.2.1: Adding gcc and gcc_s libs to link line -- Performing Test HAS_WERROR_RETURN_TYPE -- Performing Test HAS_WERROR_RETURN_TYPE - Success -- Performing Test HAS_WERROR_NON_VIRTUAL_DTOR -- Performing Test HAS_WERROR_NON_VIRTUAL_DTOR - Success -- Performing Test HAS_WERROR_BRACED_SCALAR_INIT -- Performing Test HAS_WERROR_BRACED_SCALAR_INIT - Failed -- Performing Test HAS_WERROR_RANGE_LOOP_CONSTRUCT -- Performing Test HAS_WERROR_RANGE_LOOP_CONSTRUCT - Success -- Performing Test HAS_WERROR_BOOL_OPERATION -- Performing Test HAS_WERROR_BOOL_OPERATION - Success -- Performing Test HAS_WNARROWING -- Performing Test HAS_WNARROWING - Success -- Performing Test HAS_WNO_MISSING_FIELD_INITIALIZERS -- Performing Test HAS_WNO_MISSING_FIELD_INITIALIZERS - Success -- Performing Test HAS_WNO_TYPE_LIMITS -- Performing Test HAS_WNO_TYPE_LIMITS - Success -- Performing Test HAS_WNO_ARRAY_BOUNDS -- Performing Test HAS_WNO_ARRAY_BOUNDS - Success -- Performing Test HAS_WNO_UNKNOWN_PRAGMAS -- Performing Test HAS_WNO_UNKNOWN_PRAGMAS - Success -- Performing Test HAS_WNO_UNUSED_PARAMETER -- Performing Test HAS_WNO_UNUSED_PARAMETER - Success -- Performing Test HAS_WNO_UNUSED_FUNCTION -- Performing Test HAS_WNO_UNUSED_FUNCTION - Success -- Performing Test HAS_WNO_UNUSED_RESULT -- Performing Test HAS_WNO_UNUSED_RESULT - Success -- Performing Test HAS_WNO_STRICT_OVERFLOW -- Performing Test HAS_WNO_STRICT_OVERFLOW - Success -- Performing Test HAS_WNO_STRICT_ALIASING -- Performing Test HAS_WNO_STRICT_ALIASING - Success -- Performing Test HAS_WNO_STRINGOP_OVERFLOW -- Performing Test HAS_WNO_STRINGOP_OVERFLOW - Success -- Performing Test HAS_WVLA_EXTENSION -- Performing Test HAS_WVLA_EXTENSION - Failed -- Performing Test HAS_WSUGGEST_OVERRIDE -- Performing Test HAS_WSUGGEST_OVERRIDE - Success -- Performing Test HAS_WNEWLINE_EOF -- Performing Test HAS_WNEWLINE_EOF - Failed -- Performing Test HAS_WINCONSISTENT_MISSING_OVERRIDE -- Performing Test HAS_WINCONSISTENT_MISSING_OVERRIDE - Failed -- Performing Test HAS_WINCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE -- Performing Test HAS_WINCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE - Failed CMake Warning at CMakeLists.txt:1020 (message): USE_GOLD_LINKER was set but ld.gold isn't available, turning it off -- Performing Test HAS_WNO_ERROR_PEDANTIC -- Performing Test HAS_WNO_ERROR_PEDANTIC - Success -- Performing Test HAS_WNO_ERROR_OLD_STYLE_CAST -- Performing Test HAS_WNO_ERROR_OLD_STYLE_CAST - Success -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_OVERRIDE -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_OVERRIDE - Failed -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE -- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE - Failed -- Performing Test HAS_WCONSTANT_CONVERSION -- Performing Test HAS_WCONSTANT_CONVERSION - Failed -- Performing Test HAS_WNO_INVALID_PARTIAL_SPECIALIZATION -- Performing Test HAS_WNO_INVALID_PARTIAL_SPECIALIZATION - Failed -- Performing Test HAS_WNO_ALIGNED_ALLOCATION_UNAVAILABLE -- Performing Test HAS_WNO_ALIGNED_ALLOCATION_UNAVAILABLE - Failed -- Performing Test HAS_WNO_MISSING_BRACES -- Performing Test HAS_WNO_MISSING_BRACES - Success -- Performing Test HAS_QUNUSED_ARGUMENTS -- Performing Test HAS_QUNUSED_ARGUMENTS - Failed -- Performing Test HAS_FDIAGNOSTICS_COLOR_ALWAYS -- Performing Test HAS_FDIAGNOSTICS_COLOR_ALWAYS - Success -- Performing Test HAS_FALIGNED_NEW -- Performing Test HAS_FALIGNED_NEW - Success -- Performing Test HAS_WNO_UNUSED_BUT_SET_VARIABLE -- Performing Test HAS_WNO_UNUSED_BUT_SET_VARIABLE - Success -- Performing Test HAS_WNO_MAYBE_UNINITIALIZED -- Performing Test HAS_WNO_MAYBE_UNINITIALIZED - Success -- Performing Test HAS_FSTANDALONE_DEBUG -- Performing Test HAS_FSTANDALONE_DEBUG - Failed -- Performing Test HAS_FNO_MATH_ERRNO -- Performing Test HAS_FNO_MATH_ERRNO - Success -- Performing Test HAS_FNO_TRAPPING_MATH -- Performing Test HAS_FNO_TRAPPING_MATH - Success -- Performing Test HAS_WERROR_FORMAT -- Performing Test HAS_WERROR_FORMAT - Success -- Performing Test HAS_WDEPRECATED -- Performing Test HAS_WDEPRECATED - Success -- NUMA paths: -- /usr/include -- /usr/lib64/libnuma.so -- Looking for backtrace -- Looking for backtrace - found -- backtrace facility detected in default set of libraries -- Found Backtrace: /usr/include -- headers outputs: -- sources outputs: -- declarations_yaml outputs: -- Performing Test COMPILER_SUPPORTS_NO_AVX256_SPLIT -- Performing Test COMPILER_SUPPORTS_NO_AVX256_SPLIT - Success -- Using ATen parallel backend: OMP Building PyTorch for GPU arch: gfx900;gfx906:xnack-;gfx908:xnack-;gfx90a:xnack+;gfx90a:xnack-;gfx942 HIP VERSION: 6.2.41134-0 -- Caffe2: Header version is: 6.2.0 ***** ROCm version from rocm_version.h **** ROCM_VERSION_DEV: 6.2.0 ROCM_VERSION_DEV_MAJOR: 6 ROCM_VERSION_DEV_MINOR: 2 ROCM_VERSION_DEV_PATCH: 0 ROCM_VERSION_DEV_INT: 60200 HIP_VERSION_MAJOR: 6 HIP_VERSION_MINOR: 2 TORCH_HIP_VERSION: 602 ***** Library versions from dpkg ***** ***** Library versions from cmake find_package ***** hip VERSION: 6.2.41134 hsa-runtime64 VERSION: 1.14.0 amd_comgr VERSION: 2.8.0 rocrand VERSION: 3.1.0 hiprand VERSION: 2.11.0 rocblas VERSION: 4.2.1 hipblas VERSION: 2.2.0 hipblaslt VERSION: 0.8.0 miopen VERSION: 3.2.0 hipfft VERSION: 1.0.15 hipsparse VERSION: 3.1.1 rccl VERSION: 2.20.5 rocprim VERSION: 3.2.0 hipcub VERSION: 3.2.0 rocthrust VERSION: 3.1.0 hipsolver VERSION: 2.2.0 HIP is using new type enums ROCm is enabled. Found sleef: /usr/lib64/libsleef.so AT_INSTALL_INCLUDE_DIR include/ATen/core core header install: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h core header install: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/aten_interned_strings.h core header install: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/enum_tag.h -- _GLIBCXX_USE_CXX11_ABI=1 is already defined as a cmake variable INFOcaffe2 ROCM_SOURCE_DIR = -- Using /usr/lib/python3.13/site-packages as python relative installation path CMake Warning at CMakeLists.txt:1283 (message): Generated cmake files are only fully tested if one builds with system glog, gflags, and protobuf. Other settings may generate files that are not well tested. -- -- ******** Summary ******** -- General: -- CMake version : 3.30.5 -- CMake command : /usr/bin/cmake -- System : Linux -- C++ compiler : /usr/bin/g++ -- C++ compiler id : GNU -- C++ compiler version : 14.2.1 -- Using ccache if found : ON -- Found ccache : CCACHE_PROGRAM-NOTFOUND -- CXX flags : -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-U_FORTIFY_SOURCE,-D_FORTIFY_SOURCE=3 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -mtls-dialect=gnu2 -fno-omit-frame-pointer -mno-omit-leaf-frame-pointer -D_GLIBCXX_USE_CXX11_ABI=1 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=pedantic -Wno-error=old-style-cast -Wno-missing-braces -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow -- Shared LD flags : -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -Wl,--no-as-needed -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -rdynamic -- Static LD flags : -- Module LD flags : -Wl,-z,relro -Wl,--as-needed -Wl,-z,pack-relative-relocs -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -specs=/usr/lib/rpm/redhat/redhat-package-notes -- Build type : RelWithDebInfo -- Compile definitions : ROCM_VERSION=60200;TORCH_HIP_VERSION=602;ONNX_ML=1;ONNXIFI_ENABLE_EXT=1;ONNX_NAMESPACE=onnx;HAVE_MMAP=1;_FILE_OFFSET_BITS=64;HAVE_SHM_OPEN=1;HAVE_SHM_UNLINK=1;HAVE_MALLOC_USABLE_SIZE=1;USE_EXTERNAL_MZCRC;MINIZ_DISABLE_ZIP_READER_CRC32_CHECKS;FLASHATTENTION_DISABLE_ALIBI -- CMAKE_PREFIX_PATH : /usr/lib/python3.13/site-packages -- CMAKE_INSTALL_PREFIX : /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch -- USE_GOLD_LINKER : OFF -- -- TORCH_VERSION : 2.4.0 -- BUILD_STATIC_RUNTIME_BENCHMARK: OFF -- BUILD_BINARY : OFF -- BUILD_CUSTOM_PROTOBUF : OFF -- Protobuf compiler : /usr/bin/protoc -- Protobuf includes : /usr/include -- Protobuf libraries : /usr/lib64/libprotobuf.so -- BUILD_DOCS : OFF -- BUILD_PYTHON : True -- Python version : 3.13.1 -- Python executable : /usr/bin/python3 -- Python library : -- Python includes : /usr/include/python3.13 -- Python site-package : /usr/lib/python3.13/site-packages -- BUILD_SHARED_LIBS : ON -- CAFFE2_USE_MSVC_STATIC_RUNTIME : OFF -- BUILD_TEST : False -- BUILD_JNI : OFF -- BUILD_MOBILE_AUTOGRAD : OFF -- BUILD_LITE_INTERPRETER: OFF -- INTERN_BUILD_MOBILE : -- TRACING_BASED : OFF -- USE_BLAS : 1 -- BLAS : flexi -- BLAS_HAS_SBGEMM : -- USE_LAPACK : 1 -- LAPACK : flexi -- USE_ASAN : OFF -- USE_TSAN : OFF -- USE_CPP_CODE_COVERAGE : OFF -- USE_CUDA : OFF -- USE_XPU : OFF -- USE_ROCM : ON -- ROCM_VERSION : -- USE_FLASH_ATTENTION : OFF -- USE_MEM_EFF_ATTENTION : OFF -- BUILD_NVFUSER : OFF -- USE_EIGEN_FOR_BLAS : ON -- USE_FBGEMM : OFF -- USE_FAKELOWP : OFF -- USE_KINETO : OFF -- USE_GFLAGS : OFF -- USE_GLOG : OFF -- USE_LITE_PROTO : OFF -- USE_PYTORCH_METAL : OFF -- USE_PYTORCH_METAL_EXPORT : OFF -- USE_MPS : OFF -- USE_MKL : OFF -- USE_MKLDNN : OFF -- USE_UCC : OFF -- USE_ITT : OFF -- USE_NCCL : OFF -- USE_NNPACK : OFF -- USE_NUMPY : ON -- USE_OBSERVERS : ON -- USE_OPENCL : OFF -- USE_OPENMP : ON -- USE_MIMALLOC : OFF -- USE_VULKAN : OFF -- USE_PROF : OFF -- USE_PYTORCH_QNNPACK : OFF -- USE_XNNPACK : ON -- USE_DISTRIBUTED : ON -- USE_MPI : OFF -- USE_GLOO : OFF -- USE_GLOO_WITH_OPENSSL : OFF -- USE_TENSORPIPE : ON -- Public Dependencies : -- Private Dependencies : Threads::Threads;pthreadpool;cpuinfo;XNNPACK;fp16;caffe2::openmp;tensorpipe;onnx_proto;onnx;rt;fmt;gcc_s;gcc;dl -- Public CUDA Deps. : -- Private CUDA Deps. : -- USE_COREML_DELEGATE : OFF -- BUILD_LAZY_TS_BACKEND : ON -- USE_ROCM_KERNEL_ASSERT : OFF -- Performing Test HAS_WMISSING_PROTOTYPES -- Performing Test HAS_WMISSING_PROTOTYPES - Failed -- Performing Test HAS_WERROR_MISSING_PROTOTYPES -- Performing Test HAS_WERROR_MISSING_PROTOTYPES - Failed -- Configuring done (33.5s) CMake Warning at caffe2/CMakeLists.txt:882 (add_library): Cannot generate a safe runtime search path for target torch because files in some directories may conflict with libraries in implicit directories: runtime library [libhipblas.so.2] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipfft.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhiprand.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsparse.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsolver.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib Some of these libraries may not be found correctly. CMake Warning at /usr/lib64/cmake/hip/FindHIP.cmake:764 (add_library): Cannot generate a safe runtime search path for target torch_hip because files in some directories may conflict with libraries in implicit directories: runtime library [libhipblas.so.2] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipfft.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhiprand.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsparse.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsolver.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libmagma.so.2.8.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib Some of these libraries may not be found correctly. Call Stack (most recent call first): caffe2/CMakeLists.txt:897 (hip_add_library) CMake Warning at torch/CMakeLists.txt:300 (add_library): Cannot generate a safe runtime search path for target torch_python because files in some directories may conflict with libraries in implicit directories: runtime library [libhipblas.so.2] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipfft.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhiprand.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsparse.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsolver.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib Some of these libraries may not be found correctly. CMake Warning at torch/CMakeLists.txt:404 (add_library): Cannot generate a safe runtime search path for target nnapi_backend because files in some directories may conflict with libraries in implicit directories: runtime library [libhipblas.so.2] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipfft.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhiprand.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsparse.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsolver.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib Some of these libraries may not be found correctly. CMake Warning at functorch/CMakeLists.txt:11 (add_library): Cannot generate a safe runtime search path for target functorch because files in some directories may conflict with libraries in implicit directories: runtime library [libhipblas.so.2] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipfft.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhiprand.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsparse.so.1] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib runtime library [libhipsolver.so.0] in /usr/lib64 may be hidden by files in: /usr/lib64/rocm/gfx9/lib Some of these libraries may not be found correctly. -- Generating done (0.6s) -- Build files have been written to: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build [1/4] Generating ATen headers [2/4] Generating ATen sources [3/4] Generating ATen declarations_yaml [1/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/fs-poll.c.o [2/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/idna.c.o [3/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/random.c.o [4/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/strscpy.c.o [5/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/inet.c.o [6/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/threadpool.c.o [7/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/timer.c.o [8/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/uv-data-getter-setters.c.o [9/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/version.c.o [10/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/async.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/internal.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c: In function ‘uv__async_io’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_22(D) + 432B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:149:3: note: in expansion of macro ‘QUEUE_MOVE’ 149 | QUEUE_MOVE(&loop->async_handles, &queue); | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:125:9: note: ‘queue’ declared here 125 | QUEUE queue; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/async.c:122:37: note: ‘loop’ declared here 122 | static void uv__async_io(uv_loop_t* loop, uv__io_t* w, unsigned int events) { | ~~~~~~~~~~~^~~~ [11/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/uv-common.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c: In function ‘uv_walk’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_13(D) + 16B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:491:3: note: in expansion of macro ‘QUEUE_MOVE’ 491 | QUEUE_MOVE(&loop->handle_queue, &queue); | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:487:9: note: ‘queue’ declared here 487 | QUEUE queue; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.c:486:25: note: ‘loop’ declared here 486 | void uv_walk(uv_loop_t* loop, uv_walk_cb walk_cb, void* arg) { | ~~~~~~~~~~~^~~~ [12/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/dl.c.o [13/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/core.c.o [14/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/getaddrinfo.c.o [15/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/getnameinfo.c.o [16/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/fs.c.o [17/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/loop-watcher.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/internal.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c: In function ‘uv__run_prepare’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_12(D) + 384B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:52:5: note: in expansion of macro ‘QUEUE_MOVE’ 52 | QUEUE_MOVE(&loop->name##_handles, &queue); \ | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:66:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 66 | UV_LOOP_WATCHER_DEFINE(prepare, PREPARE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:50:11: note: ‘queue’ declared here 50 | QUEUE queue; \ | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:66:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 66 | UV_LOOP_WATCHER_DEFINE(prepare, PREPARE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:48:34: note: ‘loop’ declared here 48 | void uv__run_##name(uv_loop_t* loop) { \ | ~~~~~~~~~~~^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:66:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 66 | UV_LOOP_WATCHER_DEFINE(prepare, PREPARE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c: In function ‘uv__run_check’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_12(D) + 400B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:52:5: note: in expansion of macro ‘QUEUE_MOVE’ 52 | QUEUE_MOVE(&loop->name##_handles, &queue); \ | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:67:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 67 | UV_LOOP_WATCHER_DEFINE(check, CHECK) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:50:11: note: ‘queue’ declared here 50 | QUEUE queue; \ | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:67:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 67 | UV_LOOP_WATCHER_DEFINE(check, CHECK) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:48:34: note: ‘loop’ declared here 48 | void uv__run_##name(uv_loop_t* loop) { \ | ~~~~~~~~~~~^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:67:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 67 | UV_LOOP_WATCHER_DEFINE(check, CHECK) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c: In function ‘uv__run_idle’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘queue’ in ‘((void * (**)[2])MEM[(void *[2] * *)loop_12(D) + 416B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:52:5: note: in expansion of macro ‘QUEUE_MOVE’ 52 | QUEUE_MOVE(&loop->name##_handles, &queue); \ | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:68:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 68 | UV_LOOP_WATCHER_DEFINE(idle, IDLE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:50:11: note: ‘queue’ declared here 50 | QUEUE queue; \ | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:68:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 68 | UV_LOOP_WATCHER_DEFINE(idle, IDLE) | ^~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:48:34: note: ‘loop’ declared here 48 | void uv__run_##name(uv_loop_t* loop) { \ | ~~~~~~~~~~~^~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/loop-watcher.c:68:1: note: in expansion of macro ‘UV_LOOP_WATCHER_DEFINE’ 68 | UV_LOOP_WATCHER_DEFINE(idle, IDLE) | ^~~~~~~~~~~~~~~~~~~~~~ [18/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/loop.c.o [19/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/pipe.c.o [20/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/poll.c.o [21/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/random-devurandom.c.o [22/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/process.c.o [23/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/signal.c.o [24/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/stream.c.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/uv-common.h:42, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/internal.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:23: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c: In function ‘uv__write_callbacks’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:68:19: warning: storing the address of local variable ‘pq’ in ‘((void * (**)[2])MEM[(void *[2] * *)stream_23(D) + 208B])[1]’ [-Wdangling-pointer=] 68 | QUEUE_PREV(q) = (n); \ | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/queue.h:78:7: note: in expansion of macro ‘QUEUE_SPLIT’ 78 | QUEUE_SPLIT(h, q, n); \ | ^~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:935:3: note: in expansion of macro ‘QUEUE_MOVE’ 935 | QUEUE_MOVE(&stream->write_completed_queue, &pq); | ^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:930:9: note: ‘pq’ declared here 930 | QUEUE pq; | ^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/stream.c:927:46: note: ‘stream’ declared here 927 | static void uv__write_callbacks(uv_stream_t* stream) { | ~~~~~~~~~~~~~^~~~~~ [25/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/tcp.c.o [26/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/thread.c.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c: In function ‘thread_stack_size’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c:195:24: warning: comparison of integer expressions of different signedness: ‘rlim_t’ {aka ‘long unsigned int’} and ‘long int’ [-Wsign-compare] 195 | if (lim.rlim_cur >= PTHREAD_STACK_MIN) | ^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c: In function ‘uv_thread_create_ex’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libuv/src/unix/thread.c:243:20: warning: comparison of integer expressions of different signedness: ‘size_t’ {aka ‘long unsigned int’} and ‘long int’ [-Wsign-compare] 243 | if (stack_size < PTHREAD_STACK_MIN) | ^ [27/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/tty.c.o [28/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/proctitle.c.o [29/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/udp.c.o [30/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/linux-core.c.o [31/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/linux-syscalls.c.o [32/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/procfs-exepath.c.o [33/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/linux-inotify.c.o [34/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/random-getrandom.c.o [35/1912] Building C object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe_uv.dir/__/third_party/libuv/src/unix/random-sysctl-linux.c.o [36/1912] Building CXX object c10/CMakeFiles/c10.dir/core/AutogradState.cpp.o [37/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Allocator.cpp.o [38/1912] Building CXX object c10/CMakeFiles/c10.dir/core/ConstantSymNodeImpl.cpp.o [39/1912] Building CXX object c10/CMakeFiles/c10.dir/core/CPUAllocator.cpp.o [40/1912] Building CXX object c10/CMakeFiles/c10.dir/core/CopyBytes.cpp.o [41/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DefaultDtype.cpp.o [42/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Device.cpp.o [43/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DeviceType.cpp.o [44/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DispatchKeySet.cpp.o [45/1912] Building CXX object c10/CMakeFiles/c10.dir/core/DispatchKey.cpp.o [46/1912] Building CXX object c10/CMakeFiles/c10.dir/core/GradMode.cpp.o [47/1912] Building CXX object c10/CMakeFiles/c10.dir/core/InferenceMode.cpp.o [48/1912] Building CXX object c10/CMakeFiles/c10.dir/core/GeneratorImpl.cpp.o [49/1912] Building CXX object c10/CMakeFiles/c10.dir/core/RefcountedDeleter.cpp.o [50/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SafePyObject.cpp.o [51/1912] Building CXX object c10/CMakeFiles/c10.dir/core/ScalarType.cpp.o [52/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Scalar.cpp.o [53/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Storage.cpp.o [54/1912] Building CXX object c10/CMakeFiles/c10.dir/core/StorageImpl.cpp.o [55/1912] Building CXX object c10/CMakeFiles/c10.dir/core/Stream.cpp.o [56/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymBool.cpp.o [57/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymFloat.cpp.o [58/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymIntArrayRef.cpp.o [59/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymNodeImpl.cpp.o [60/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymInt.cpp.o [61/1912] Building CXX object c10/CMakeFiles/c10.dir/core/SymbolicShapeMeta.cpp.o [62/1912] Building CXX object c10/CMakeFiles/c10.dir/core/TensorOptions.cpp.o [63/1912] Building CXX object c10/CMakeFiles/c10.dir/core/UndefinedTensorImpl.cpp.o [64/1912] Building CXX object c10/CMakeFiles/c10.dir/core/WrapDimMinimal.cpp.o [65/1912] Building CXX object c10/CMakeFiles/c10.dir/core/TensorImpl.cpp.o [66/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/COWDeleter.cpp.o [67/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/COW.cpp.o [68/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/DeviceGuardImplInterface.cpp.o [69/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/HermeticPyObjectTLS.cpp.o [70/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/GPUTrace.cpp.o [71/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/LocalDispatchKeySet.cpp.o [72/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/PyInterpreter.cpp.o [73/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/PyObjectSlot.cpp.o [74/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/PythonDispatcherTLS.cpp.o [75/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/SizesAndStrides.cpp.o [76/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/TorchDispatchModeTLS.cpp.o [77/1912] Building CXX object c10/CMakeFiles/c10.dir/core/impl/alloc_cpu.cpp.o [78/1912] Building CXX object c10/CMakeFiles/c10.dir/core/thread_pool.cpp.o [79/1912] Building CXX object c10/CMakeFiles/c10.dir/mobile/CPUCachingAllocator.cpp.o [80/1912] Building CXX object c10/CMakeFiles/c10.dir/util/ApproximateClock.cpp.o [81/1912] Building CXX object c10/CMakeFiles/c10.dir/mobile/CPUProfilingAllocator.cpp.o [82/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Bfloat16.cpp.o [83/1912] Building CXX object c10/CMakeFiles/c10.dir/util/C++17.cpp.o [84/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Backtrace.cpp.o [85/1912] Building CXX object c10/CMakeFiles/c10.dir/util/DeadlockDetection.cpp.o [86/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e4m3fn.cpp.o [87/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e4m3fnuz.cpp.o [88/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e5m2.cpp.o [89/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Exception.cpp.o [90/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Float8_e5m2fnuz.cpp.o [91/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Half.cpp.o [92/1912] Building CXX object c10/CMakeFiles/c10.dir/util/LeftRight.cpp.o [93/1912] Building CXX object c10/CMakeFiles/c10.dir/util/MathConstants.cpp.o [94/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Metaprogramming.cpp.o [95/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Optional.cpp.o [96/1912] Building CXX object c10/CMakeFiles/c10.dir/util/ParallelGuard.cpp.o [97/1912] Building CXX object c10/CMakeFiles/c10.dir/util/SmallVector.cpp.o [98/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Logging.cpp.o [99/1912] Building CXX object c10/CMakeFiles/c10.dir/util/StringUtil.cpp.o [100/1912] Building CXX object c10/CMakeFiles/c10.dir/util/TypeCast.cpp.o [101/1912] Building CXX object c10/CMakeFiles/c10.dir/util/ThreadLocalDebugInfo.cpp.o [102/1912] Building CXX object c10/CMakeFiles/c10.dir/util/TypeList.cpp.o [103/1912] Building CXX object c10/CMakeFiles/c10.dir/util/TypeTraits.cpp.o [104/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Type_no_demangle.cpp.o [105/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Unicode.cpp.o [106/1912] Building CXX object c10/CMakeFiles/c10.dir/util/UniqueVoidPtr.cpp.o [107/1912] Building CXX object c10/CMakeFiles/c10.dir/util/Type_demangle.cpp.o [108/1912] Building CXX object c10/CMakeFiles/c10.dir/util/flags_use_gflags.cpp.o [109/1912] Building CXX object c10/CMakeFiles/c10.dir/util/complex_math.cpp.o [110/1912] Building CXX object c10/CMakeFiles/c10.dir/util/int128.cpp.o [111/1912] Building CXX object c10/CMakeFiles/c10.dir/util/flags_use_no_gflags.cpp.o [112/1912] Building CXX object c10/CMakeFiles/c10.dir/util/intrusive_ptr.cpp.o [113/1912] Building CXX object c10/CMakeFiles/c10.dir/util/numa.cpp.o [114/1912] Building CXX object c10/CMakeFiles/c10.dir/util/signal_handler.cpp.o [115/1912] Building CXX object c10/CMakeFiles/c10.dir/util/thread_name.cpp.o [116/1912] Building CXX object c10/CMakeFiles/c10.dir/util/tempfile.cpp.o [117/1912] Linking C static library lib/libtensorpipe_uv.a [118/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/error.cc.o [119/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/helpers.cc.o [120/1912] Building CXX object c10/CMakeFiles/c10.dir/util/typeid.cpp.o [121/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/address.cc.o [122/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/allocator.cc.o [123/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/error.cc.o [124/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/fd.cc.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc: In member function ‘tensorpipe::Error tensorpipe::Fd::readFull(void*, size_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc:50:10: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 50 | if (rv != count) { | ~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc: In member function ‘tensorpipe::Error tensorpipe::Fd::writeFull(const void*, size_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/fd.cc:62:10: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 62 | if (rv != count) { | ~~~^~~~~~~~ [125/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/socket.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [126/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/system.cc.o [127/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/context.cc.o [128/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/error.cc.o [129/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/listener.cc.o [130/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/context_impl.cc.o [131/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/pipe.cc.o [132/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/listener_impl.cc.o [133/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/error.cc.o [134/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/basic/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::SendOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.cc:51:28: required from here 51 | sendOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::basic::ChannelImpl; TOp = tensorpipe::channel::basic::RecvOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/basic/channel_impl.cc:118:28: required from here 118 | recvOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [135/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/basic/context_impl.cc.o [136/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/basic/factory.cc.o [137/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/core/pipe_impl.cc.o /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc: In function ‘void tensorpipe::{anonymous}::parseDescriptorReplyOfMessage(tensorpipe::WriteOperation&, tensorpipe::DescriptorReply)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:35:40: warning: comparison of integer expressions of different signedness: ‘size_t’ {aka ‘long unsigned int’} and ‘const int’ [-Wsign-compare] 35 | for (size_t tensorIdx = 0; tensorIdx < numTensors; ++tensorIdx) { | ~~~~~~~~~~^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc: In function ‘std::shared_ptr > tensorpipe::{anonymous}::makeDescriptorForMessage(const tensorpipe::WriteOperation&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:77:39: warning: comparison of integer expressions of different signedness: ‘int’ and ‘std::vector::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 77 | for (int payloadIdx = 0; payloadIdx < op.message.payloads.size(); | ~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:87:37: warning: comparison of integer expressions of different signedness: ‘int’ and ‘std::vector::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 87 | for (int tensorIdx = 0; tensorIdx < op.tensors.size(); ++tensorIdx) { | ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc: In member function ‘void tensorpipe::PipeImpl::expectReadCall(ReadOpIter)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:869:18: warning: unused variable ‘op’ [-Wunused-variable] 869 | ReadOperation& op = *opIter; | ^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.h:25, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::ReadOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::ReadOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:391:28: required from here 391 | readOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.h:21: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::WriteOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::PipeImpl; TOp = tensorpipe::WriteOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.cc:580:29: required from here 580 | writeOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/serializer.h:24, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/nop_types.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/core/pipe_impl.h:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::pair; T = std::__cxx11::basic_string; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::__cxx11::basic_string > >; Type = std::unordered_map, std::__cxx11::basic_string >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::__cxx11::basic_string >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transport, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::address, void*, 0, void>, nop::MemberPointer, std::equal_to, std::allocator > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportDomainDescriptor, void*, 0, void>, nop::MemberPointer, std::allocator >, std::vector >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::vector > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelDeviceDescriptors, void*, 0, void>, nop::MemberPointer, std::__cxx11::basic_string, std::allocator >, std::hash >, std::equal_to >, std::allocator, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelForDevicePair, void*, 0, void> >; T = std::unordered_map, std::__cxx11::basic_string >; Class = tensorpipe::BrochureAnswer; T Class::* Pointer = &tensorpipe::BrochureAnswer::channelForDevicePair]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 7; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::BrochureAnswer]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::__cxx11::basic_string >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, false>::value_type’ {aka ‘const std::pair, std::__cxx11::basic_string >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::__cxx11::basic_string >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, false>::value_type&’ {aka ‘const std::pair, std::__cxx11::basic_string >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::unordered_map >; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::unordered_map > > >; Type = std::unordered_map, std::unordered_map > >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::unordered_map > >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::transportDomainDescriptors, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::channelDeviceDescriptors, void*, 0, void> >; T = std::unordered_map, std::unordered_map > >; Class = tensorpipe::Brochure; T Class::* Pointer = &tensorpipe::Brochure::channelDeviceDescriptors]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 2; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::Brochure]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::unordered_map > >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::unordered_map > >, false, true>::value_type’ {aka ‘const std::pair, std::unordered_map > >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::unordered_map > >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::unordered_map > >, false, true>::value_type&’ {aka ‘const std::pair, std::unordered_map > >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::__cxx11::basic_string; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::__cxx11::basic_string > >; Type = std::unordered_map, std::__cxx11::basic_string >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::__cxx11::basic_string >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::transportDomainDescriptors, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::channelDeviceDescriptors, void*, 0, void> >; T = std::unordered_map, std::__cxx11::basic_string >; Class = tensorpipe::Brochure; T Class::* Pointer = &tensorpipe::Brochure::transportDomainDescriptors]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 1; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 2; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 99 | auto status = WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::Brochure]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::__cxx11::basic_string >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, true>::value_type’ {aka ‘const std::pair, std::__cxx11::basic_string >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::__cxx11::basic_string >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::__cxx11::basic_string >, false, true>::value_type&’ {aka ‘const std::pair, std::__cxx11::basic_string >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::vector; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::vector > >; Type = std::unordered_map, std::vector >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::vector >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transport, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::address, void*, 0, void>, nop::MemberPointer, std::equal_to, std::allocator > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportDomainDescriptor, void*, 0, void>, nop::MemberPointer, std::allocator >, std::vector >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::vector > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelDeviceDescriptors, void*, 0, void>, nop::MemberPointer, std::__cxx11::basic_string, std::allocator >, std::hash >, std::equal_to >, std::allocator, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelForDevicePair, void*, 0, void> >; T = std::unordered_map, std::vector >; Class = tensorpipe::BrochureAnswer; T Class::* Pointer = &tensorpipe::BrochureAnswer::channelRegistrationIds]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: recursively required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 6; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 99 | auto status = WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 7; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::BrochureAnswer]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair, std::vector >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, std::vector >, false, true>::value_type’ {aka ‘const std::pair, std::vector >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair, std::vector >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, std::vector >, false, true>::value_type&’ {aka ‘const std::pair, std::vector >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = tensorpipe::Device; T = std::__cxx11::basic_string; Hash = std::hash; KeyEqual = std::equal_to; Allocator = std::allocator > >; Type = std::unordered_map >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:147:34: required from ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = std::__cxx11::basic_string; T = std::unordered_map >; Hash = std::hash >; KeyEqual = std::equal_to >; Allocator = std::allocator, std::unordered_map > > >; Type = std::unordered_map, std::unordered_map > >]’ 147 | status = Encoding::Write(element.second, writer); | ~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map, std::unordered_map > >]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator >, std::__cxx11::basic_string, std::allocator >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::transportDomainDescriptors, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::Brochure::*, &tensorpipe::Brochure::channelDeviceDescriptors, void*, 0, void> >; T = std::unordered_map, std::unordered_map > >; Class = tensorpipe::Brochure; T Class::* Pointer = &tensorpipe::Brochure::channelDeviceDescriptors]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:103:41: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 2; Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 103 | return PointerAt::Write(value, writer, MemberList{}); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::Brochure]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::Brochure]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair >&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator >, false, false>::value_type’ {aka ‘const std::pair >’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair >’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator >, false, false>::value_type&’ {aka ‘const std::pair >&’} to prevent copying /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h: In instantiation of ‘static constexpr nop::Status nop::Encoding >::WritePayload(nop::EncodingByte, const Type&, Writer*) [with Writer = tensorpipe::NopWriter; Key = long unsigned int; T = long unsigned int; Hash = std::hash; KeyEqual = std::equal_to; Allocator = std::allocator >; Type = std::unordered_map]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = std::unordered_map]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/types/detail/member_pointer.h:56:33: required from ‘static constexpr nop::Status nop::MemberPointer::Write(const Class&, Writer*, MemberList) [with Writer = tensorpipe::NopWriter; MemberList = nop::MemberList, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transport, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::address, void*, 0, void>, nop::MemberPointer, std::equal_to, std::allocator > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::transportDomainDescriptor, void*, 0, void>, nop::MemberPointer, std::allocator >, std::vector >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::vector > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelRegistrationIds, void*, 0, void>, nop::MemberPointer, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > >, std::hash, std::allocator > >, std::equal_to, std::allocator > >, std::allocator, std::allocator >, std::unordered_map, std::allocator >, std::hash, std::equal_to, std::allocator, std::allocator > > > > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelDeviceDescriptors, void*, 0, void>, nop::MemberPointer, std::__cxx11::basic_string, std::allocator >, std::hash >, std::equal_to >, std::allocator, std::__cxx11::basic_string, std::allocator > > > > tensorpipe::BrochureAnswer::*, &tensorpipe::BrochureAnswer::channelForDevicePair, void*, 0, void> >; T = std::unordered_map; Class = tensorpipe::BrochureAnswer; T Class::* Pointer = &tensorpipe::BrochureAnswer::transportRegistrationIds]’ 56 | return Encoding::Write(Resolve(instance), writer); | ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: recursively required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 6; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 99 | auto status = WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:99:31: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WriteMembers(const T&, Writer*, nop::Index) [with long unsigned int index = 7; Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/members.h:59:26: required from ‘static constexpr nop::Status nop::Encoding::value, void>::type>::WritePayload(nop::EncodingByte, const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 59 | return WriteMembers(value, writer, Index{}); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/encoding.h:120:39: required from ‘static constexpr nop::Status nop::EncodingIO::Write(const T&, Writer*) [with Writer = tensorpipe::NopWriter; T = tensorpipe::BrochureAnswer]’ 120 | return Encoding::WritePayload(prefix, value, writer); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:232:35: required from ‘nop::Status tensorpipe::NopHolder::write(tensorpipe::NopWriter&) const [with T = tensorpipe::BrochureAnswer]’ 232 | return nop::Encoding::Write(object_, &writer); | ~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/nop.h:231:21: required from here 231 | nop::Status write(NopWriter& writer) const override { | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: warning: loop variable ‘element’ of type ‘const std::pair&’ binds to a temporary constructed from type ‘const std::__detail::_Node_const_iterator, false, false>::value_type’ {aka ‘const std::pair’} [-Wrange-loop-construct] 142 | for (const std::pair& element : value) { | ^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/third_party/libnop/include/nop/base/map.h:142:35: note: use non-reference type ‘const std::pair’ to make the copy explicit or ‘const std::__detail::_Node_const_iterator, false, false>::value_type&’ {aka ‘const std::pair&’} to prevent copying [138/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/xth/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::SendOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.cc:65:28: required from here 65 | sendOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::xth::ChannelImpl; TOp = tensorpipe::channel::xth::RecvOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/channel_impl.cc:157:28: required from here 157 | recvOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [139/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/xth/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/context_impl.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/context_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h: In instantiation of ‘void tensorpipe::Queue::push(T) [with T = tensorpipe::optional]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/xth/context_impl.cc:68:17: required from here 68 | requests_.push(nullopt); | ~~~~~~~~~~~~~~^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h:24:26: warning: comparison of integer expressions of different signedness: ‘std::deque, std::allocator > >::size_type’ {aka ‘long unsigned int’} and ‘const int’ [-Wsign-compare] 24 | while (items_.size() >= capacity_) { | ~~~~~~~~~~~~~~^~~~~~~~~~~~ [140/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/xth/factory.cc.o [141/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/cma/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.h:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::SendOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.cc:66:28: required from here 66 | sendOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:57:20: required from ‘void tensorpipe::OpsStateMachine::advanceOperation(Iter) [with TSubject = tensorpipe::channel::cma::ChannelImpl; TOp = tensorpipe::channel::cma::RecvOperation]’ 57 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/channel_impl.cc:160:28: required from here 160 | recvOps_.advanceOperation(opIter); | ~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [142/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/cma/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:26: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/strings.h: In function ‘bool tensorpipe::isValidUuid(const std::string&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/strings.h:58:21: warning: comparison of integer expressions of different signedness: ‘int’ and ‘std::__cxx11::basic_string::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 58 | for (int i = 0; i < uuid.size(); i++) { | ~~^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc: In function ‘tensorpipe::Error tensorpipe::channel::cma::{anonymous}::callProcessVmReadv(void*, void*, size_t, pid_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:61:20: warning: comparison of integer expressions of different signedness: ‘ssize_t’ {aka ‘long int’} and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 61 | } else if (nread != length) { | ~~~~~~^~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h: In instantiation of ‘void tensorpipe::Queue::push(T) [with T = tensorpipe::optional]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/cma/context_impl.cc:324:17: required from here 324 | requests_.push(nullopt); | ~~~~~~~~~~~~~~^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/queue.h:24:26: warning: comparison of integer expressions of different signedness: ‘std::deque, std::allocator > >::size_type’ {aka ‘long unsigned int’} and ‘const int’ [-Wsign-compare] 24 | while (items_.size() >= capacity_) { | ~~~~~~~~~~~~~~^~~~~~~~~~~~ [143/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/cma/factory.cc.o [144/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/mpt/channel_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::SendOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:74:20: required from ‘void tensorpipe::OpsStateMachine::advanceAllOperations() [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::SendOperation]’ 74 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.cc:131:32: required from here 131 | sendOps_.advanceAllOperations(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/callback.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/channel_impl_boilerplate.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.h:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h: In instantiation of ‘TOp* tensorpipe::OpsStateMachine::findOperation(int64_t) [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::RecvOperation; int64_t = long int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:74:20: required from ‘void tensorpipe::OpsStateMachine::advanceAllOperations() [with TSubject = tensorpipe::channel::mpt::ChannelImpl; TOp = tensorpipe::channel::mpt::RecvOperation]’ 74 | TOp* opPtr = findOperation(sequenceNumber); | ^~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/channel/mpt/channel_impl.cc:132:32: required from here 132 | recvOps_.advanceAllOperations(); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:102:30: warning: comparison of integer expressions of different signedness: ‘int64_t’ {aka ‘long int’} and ‘std::deque >::size_type’ {aka ‘long unsigned int’} [-Wsign-compare] 102 | if (offset < 0 || offset >= ops_.size()) { | ~~~~~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘uint64_t’ {aka ‘long unsigned int’} and ‘int64_t’ {aka ‘long int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/state_machine.h:106:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 106 | TP_DCHECK_EQ(op.sequenceNumber, sequenceNumber); | ^~~~~~~~~~~~ [145/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/mpt/context_impl.cc.o [146/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/channel/mpt/factory.cc.o [147/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/uv.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/loop.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/context_impl.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/context_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [148/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/connection_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/connection_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/connection_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [149/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/error.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/uv.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/error.cc:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [150/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/listener_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/listener_impl.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/listener_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [151/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/factory.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/connection_impl.h:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/factory.cc:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [152/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/loop.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/uv.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/loop.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/loop.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [153/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/sockaddr.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [154/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/uv/utility.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/uv/utility.cc:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [155/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/epoll_loop.cc.o [156/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/shm_segment.cc.o [157/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/context_impl.cc:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [158/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/connection_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘tensorpipe::Error tensorpipe::recvFromSocket(int, T&, T&, Fds& ...) [with T = unsigned int; Fds = {Fd, Fd, Fd, Fd}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:253:26: required from ‘tensorpipe::Error tensorpipe::Socket::recvPayloadAndFds(T&, T&, Fds& ...) [with T = unsigned int; Fds = {tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd}; typename std::enable_if::value, bool>::type = false]’ 253 | return recvFromSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:219:41: required from here 219 | auto err = socket_.recvPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 220 | peerInboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~ 221 | peerOutboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~~ 222 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 223 | reactorDataFd, | ~~~~~~~~~~~~~~ 224 | outboxHeaderFd, | ~~~~~~~~~~~~~~~ 225 | outboxDataFd); | ~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:154:12: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 154 | if (rv != iov.iov_len) { | ~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘tensorpipe::Error tensorpipe::sendToSocket(int, const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:243:24: required from ‘tensorpipe::Error tensorpipe::Socket::sendPayloadAndFds(const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}; typename std::enable_if::value, bool>::type = false]’ 243 | return sendToSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:274:41: required from here 274 | auto err = socket_.sendPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 275 | inboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ 276 | outboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 277 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 278 | reactorDataFd, | ~~~~~~~~~~~~~~ 279 | inboxHeaderSegment_.getFd(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 280 | inboxDataSegment_.getFd()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:107:12: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 107 | if (rv != iov.iov_len) { | ~~~^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘void tensorpipe::{anonymous}::loadFdsFromArray(int*, std::index_sequence, Fds& ...) [with long unsigned int ...Idxs = {0, 1, 2, 3}; Fds = {tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd}; std::index_sequence = std::integer_sequence]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:171:19: required from ‘tensorpipe::Error tensorpipe::recvFromSocket(int, T&, T&, Fds& ...) [with T = unsigned int; Fds = {Fd, Fd, Fd, Fd}]’ 171 | loadFdsFromArray(payload, std::index_sequence_for{}, fds...); | ~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:253:26: required from ‘tensorpipe::Error tensorpipe::Socket::recvPayloadAndFds(T&, T&, Fds& ...) [with T = unsigned int; Fds = {tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd, tensorpipe::Fd}; typename std::enable_if::value, bool>::type = false]’ 253 | return recvFromSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:219:41: required from here 219 | auto err = socket_.recvPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 220 | peerInboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~ 221 | peerOutboxReactorToken, | ~~~~~~~~~~~~~~~~~~~~~~~ 222 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 223 | reactorDataFd, | ~~~~~~~~~~~~~~ 224 | outboxHeaderFd, | ~~~~~~~~~~~~~~~ 225 | outboxDataFd); | ~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:58:8: warning: unused variable ‘dummy’ [-Wunused-variable] 58 | auto dummy = {(loadOneFdFromArray(array[Idxs], fds), 0)...}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h: In instantiation of ‘void tensorpipe::{anonymous}::saveFdsToArray(int*, std::index_sequence, const Fds& ...) [with long unsigned int ...Idxs = {0, 1, 2, 3}; Fds = {int, int, int, int}; std::index_sequence = std::integer_sequence]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:96:17: required from ‘tensorpipe::Error tensorpipe::sendToSocket(int, const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}]’ 96 | saveFdsToArray(payload, std::index_sequence_for{}, fds...); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:243:24: required from ‘tensorpipe::Error tensorpipe::Socket::sendPayloadAndFds(const T&, const T&, const Fds& ...) [with T = unsigned int; Fds = {int, int, int, int}; typename std::enable_if::value, bool>::type = false]’ 243 | return sendToSocket(fd_, t1, t2, fds...); | ~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.cc:274:41: required from here 274 | auto err = socket_.sendPayloadAndFds( | ~~~~~~~~~~~~~~~~~~~~~~~~~^ 275 | inboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ 276 | outboxReactorToken_.value(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 277 | reactorHeaderFd, | ~~~~~~~~~~~~~~~~ 278 | reactorDataFd, | ~~~~~~~~~~~~~~ 279 | inboxHeaderSegment_.getFd(), | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 280 | inboxDataSegment_.getFd()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:41:8: warning: unused variable ‘dummy’ [-Wunused-variable] 41 | auto dummy = {(saveOneFdToArray(array[Idxs], fds), 0)...}; | ^~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ [159/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/factory.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/connection_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/factory.cc:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [160/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/listener_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/listener_impl.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/listener_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [161/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/sockaddr.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/sockaddr.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/sockaddr.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [162/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/common/ibv.cc.o [163/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/shm/reactor.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/deferred_executor.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/busy_polling_loop.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.h:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h: In instantiation of ‘ssize_t tensorpipe::RingBufferRole::write(const void*, size_t) [with int NumRoles = 2; int RoleIdx = 1; ssize_t = long int; size_t = long unsigned int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.cc:22:29: required from here 22 | auto rv = producer.write(&token, sizeof(token)); | ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘size_t’ {aka ‘long unsigned int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h:270:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 270 | TP_DCHECK_EQ(ret, size); | ^~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h: In instantiation of ‘ssize_t tensorpipe::RingBufferRole::read(void*, size_t) [with int NumRoles = 2; int RoleIdx = 0; ssize_t = long int; size_t = long unsigned int]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/shm/reactor.cc:114:34: required from here 114 | auto ret = reactorConsumer.read(&token, sizeof(token)); | ~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:49: warning: comparison of integer expressions of different signedness: ‘long int’ and ‘const size_t’ {aka ‘const long unsigned int’} [-Wsign-compare] 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:27:39: note: in definition of macro ‘unlikely’ 27 | #define unlikely(x) __builtin_expect((x) ? 1 : 0, 0) | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:226:3: note: in expansion of macro ‘__TP_DCHECK_CMP’ 226 | __TP_DCHECK_CMP(a, b, op) | ^~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/defs.h:243:28: note: in expansion of macro ‘_TP_DCHECK_CMP’ 243 | #define TP_DCHECK_EQ(a, b) _TP_DCHECK_CMP(a, b, ==) | ^~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/ringbuffer_role.h:248:5: note: in expansion of macro ‘TP_DCHECK_EQ’ 248 | TP_DCHECK_EQ(ret, size); | ^~~~~~~~~~~~ [164/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/context_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/context_impl.cc:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [165/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/connection_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [166/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/error.cc.o [167/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/listener_impl.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/listener_impl.h:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/listener_impl.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [168/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/factory.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/connection_impl.h:23, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/factory.cc:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [169/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/sockaddr.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/sockaddr.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/sockaddr.cc:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [170/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/reactor.cc.o [171/1912] Building CXX object third_party/tensorpipe/tensorpipe/CMakeFiles/tensorpipe.dir/transport/ibv/utility.cc.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/transport/ibv/utility.cc:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:48:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, tensorpipe::Fd&)’ defined but not used [-Wunused-function] 48 | void loadOneFdFromArray(int& src, Fd& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:44:6: warning: ‘void tensorpipe::{anonymous}::loadOneFdFromArray(int&, int&)’ defined but not used [-Wunused-function] 44 | void loadOneFdFromArray(int& src, int& dst) { | ^~~~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:31:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const tensorpipe::Fd&)’ defined but not used [-Wunused-function] 31 | void saveOneFdToArray(int& dst, const Fd& src) { | ^~~~~~~~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/third_party/tensorpipe/tensorpipe/common/socket.h:27:6: warning: ‘void tensorpipe::{anonymous}::saveOneFdToArray(int&, const int&)’ defined but not used [-Wunused-function] 27 | void saveOneFdToArray(int& dst, const int& src) { | ^~~~~~~~~~~~~~~~ [172/1912] Linking CXX shared library lib/libc10.so [173/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx.dir/common_avx.cc.o [174/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx.dir/typed_axpy_avx.cc.o [175/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/adagrad_avx2.cc.o [176/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/batch_box_cox_avx2.cc.o [177/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/common_avx2.cc.o [178/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_avx2.cc.o [179/1912] Generating /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/Functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_3.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType_4.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_3.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/TraceType_4.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ADInplaceOrViewType_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ADInplaceOrViewType_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cpu.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNativeFunctions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterAutogradLazy.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/RegisterLazy.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/Functions.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/variable_factories.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/ViewFuncs.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/VariableType.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyIr.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNonNativeIr.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/lazy/generated/LazyNativeFunctions.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_3.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions_4.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_variable_methods.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_0.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_1.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_torch_functions_2.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_nn_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_fft_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_linalg_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_nested_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_sparse_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_special_functions.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_return_types.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_enum_tag.cpp, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_functions.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/autograd/generated/python_return_types.h, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/testing/_internal/generated/annotated_fn_args.py, /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/inductor/aoti_torch/generated/c_shim_cuda.cpp [180/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_fused_8bit_rowwise_avx2.cc.o [181/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_fused_8bit_rowwise_idx_avx2.cc.o [182/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/lstm_unit_cpu_avx2.cc.o [183/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/math_cpu_avx2.cc.o [184/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/embedding_lookup_idx_avx2.cc.o [185/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx2.dir/typed_axpy_avx2.cc.o [186/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx512.dir/common_avx512.cc.o [187/1912] Linking C static library lib/libfp16.a [188/1912] Linking CXX static library lib/libtensorpipe.a [189/1912] Building CXX object caffe2/perfkernels/CMakeFiles/Caffe2_perfkernels_avx512.dir/adagrad_avx512.cc.o [190/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPAllocatorConfig.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [191/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPDeviceAssertionHost.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [192/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPException.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [193/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPFunctions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [194/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPMallocAsyncAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [195/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPMiscFunctions.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [196/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPStream.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [197/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/driver_api.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [198/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/impl/HIPGuardImpl.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [199/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/impl/HIPTest.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [200/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/AccumulateType.cpp.o [201/1912] Building CXX object c10/hip/CMakeFiles/c10_hip.dir/HIPCachingAllocator.cpp.o cc1plus: warning: command-line option ‘-Wno-duplicate-decl-specifier’ is valid for C/ObjC but not for C++ [202/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/CPUGeneratorImpl.cpp.o [203/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ConjugateFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MathBitsFallback.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ConjugateFallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [204/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Context.cpp.o [205/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/CachedTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CachedTensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [206/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/DeviceAccelerator.cpp.o [207/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Dispatch.cpp.o [208/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/DynamicLibrary.cpp.o [209/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/EmptyTensor.cpp.o [210/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/DLConvertor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DLConvertor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DLConvertor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [211/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FuncTorchTLS.cpp.o [212/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ExpandUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [213/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalStorageImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalStorageImpl.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [214/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalInverses.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalInverses.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [215/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalTensorWrapper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [216/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/FunctionalizeFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalizeFallbackKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [217/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyBatchedTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedTensorImpl.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [218/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyBatchedFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedFallback.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchedFallback.h:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [219/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyVmapMode.cpp.o [220/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyVmapTransforms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyVmapTransforms.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyVmapTransforms.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [221/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/MapAllocator.cpp.o [222/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/MemoryOverlap.cpp.o [223/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/NamedTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [224/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/LegacyBatchingRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/LegacyBatchingRegistrations.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [225/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelCommon.cpp.o [226/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelNative.cpp.o [227/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelOpenMP.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelFuture.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelOpenMP.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [228/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/NestedTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NestedTensorImpl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [229/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/PythonTorchFunctionTLS.cpp.o [230/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SavedTensorHooks.cpp.o [231/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ParallelThreadPoolNative.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ParallelThreadPoolNative.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [232/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SequenceNumber.cpp.o [233/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ScalarOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [234/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SparseCsrTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorImpl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [235/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/SparseTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [236/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorGeometry.cpp.o [237/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/StorageUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/StorageUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [238/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorIndexing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [239/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorMeta.cpp.o [240/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorNames.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [241/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorIterator.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIterator.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [242/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ThreadLocalPythonObjects.cpp.o [243/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ThreadLocalState.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ThreadLocalState.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [244/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/TensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [245/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Version.cpp.o [246/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Utils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Utils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [247/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/VmapModeRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/VmapModeRegistrations.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [248/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/ZeroTensorFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp: In function ‘void at::zeroTensorFallback(const c10::OperatorHandle&, c10::DispatchKeySet, torch::jit::Stack*)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:69:25: warning: possibly dangling reference to a temporary [-Wdangling-reference] 69 | const Tensor& tensor = tensors[j]; | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ZeroTensorFallback.cpp:69:43: note: the temporary was destroyed at the end of the full expression ‘tensors.c10::List::operator[](((long unsigned int)j)).c10::impl::ListElementReference > >::operator std::conditional_t()’ 69 | const Tensor& tensor = tensors[j]; | ^ [249/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/cpu/FlushDenormal.cpp.o [250/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/cpu/Utils.cpp.o [251/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/CPUGuardImpl.cpp.o [252/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/CUDAHooksInterface.cpp.o [253/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/HIPHooksInterface.cpp.o [254/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/IPUHooksInterface.cpp.o [255/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MAIAHooksInterface.cpp.o [256/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MPSHooksInterface.cpp.o [257/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MTIAHooksInterface.cpp.o [258/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/MetaGuardImpl.cpp.o [259/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/PrivateUse1HooksInterface.cpp.o [260/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/detail/XPUHooksInterface.cpp.o [261/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/ADInterpreters.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/ADInterpreters.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/ADInterpreters.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [262/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/autocast_mode.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/autocast_mode.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [263/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesActivation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesActivation.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [264/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesBinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesBinaryOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [265/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesConvolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesConvolution.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [266/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesDynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesDynamic.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesDynamic.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [267/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesDecompositions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesDecompositions.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [268/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesFactory.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesFactory.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [269/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesHelper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [270/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesLoss.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesLoss.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [271/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesLinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesLinearAlgebra.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [272/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesModules.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesModules.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [273/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesNorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesNorm.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [274/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesPooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesPooling.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [275/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesRandomness.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesRandomness.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesRandomness.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [276/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesReduceOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesReduceOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [277/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesScatterOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesScatterOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [278/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesUnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesUnaryOps.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [279/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchRulesViews.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesViews.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchRulesHelper.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [280/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchedTensorImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedTensorImpl.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [281/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/BatchedFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedFallback.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedFallback.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/BatchedFallback.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [282/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/FunctionalizeInterpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/FunctionalizeInterpreter.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/FunctionalizeInterpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [283/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/Interpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [284/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/DynamicLayer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [285/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/LegacyVmapTransforms.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/LegacyVmapTransforms.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [286/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/LegacyBatchingRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/LegacyBatchingRegistrations.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [287/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/PlumbingHelper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/TensorWrapper.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/PlumbingHelper.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [288/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/TensorWrapper.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/TensorWrapper.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/TensorWrapper.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [289/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/VmapInterpreter.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/VmapInterpreter.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/VmapInterpreter.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [290/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/PyTorchOperatorHacks.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/PyTorchOperatorHacks.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/Interpreter.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/DynamicLayer.h:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [291/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/record_function.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/record_function.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [292/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/ATenGeneral.cpp.o [293/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/BackendSelectFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/BackendSelectFallbackKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [294/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/functorch/VmapModeRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/functorch/VmapModeRegistrations.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [295/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/DeprecatedTypePropertiesRegistry.cpp.o [296/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/DeprecatedTypeProperties.cpp.o [297/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Dimname.cpp.o [298/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Dict.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [299/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Formatting.cpp.o [300/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Generator.cpp.o [301/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/GeneratorForPrivateuseone.cpp.o [302/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/List.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [303/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/MetaFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/MetaFallbackKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [304/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/NamedTensor.cpp.o [305/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/NestedIntSymNodeImpl.cpp.o [306/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/PythonFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonFallbackKernel.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonFallbackKernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [307/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/NamedRegistrations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/NamedRegistrations.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [308/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Range.cpp.o [309/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/PythonOpRegistrationTrampoline.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonOpRegistrationTrampoline.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/PythonOpRegistrationTrampoline.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [310/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Tensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [311/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/TorchDispatchUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TorchDispatchUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [312/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/VariableHooksInterface.cpp.o [313/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/Vitals.cpp.o [314/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/VariableFallbackKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/VariableFallbackKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [315/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/blob.cpp.o [316/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/adaption.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/adaption.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/adaption.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [317/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/boxing/KernelFunction.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [318/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/class_type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [319/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/custom_class.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/custom_class.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/custom_class.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [320/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/DispatchKeyExtractor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [321/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/ObservedOperators.cpp.o [322/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/Dispatcher.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [323/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dispatch/OperatorEntry.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [324/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/function_schema.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [325/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/dynamic_type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/class_type.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dynamic_type.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [326/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/interned_strings.cpp.o [327/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/library.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/library.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [328/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/ivalue.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [329/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/op_registration/infer_schema.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [330/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/operator_name.cpp.o [331/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/register_symbols.cpp.o [332/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/op_registration/op_registration.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/op_registration.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [333/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/tensor_type.cpp.o [334/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/type_factory.cpp.o [335/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [336/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/error_report.cpp.o [337/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/union_type.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict_inl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Dict.h:397, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/union_type.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [338/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/lexer.cpp.o [339/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/function_schema_parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/function_schema_parser.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [340/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/strtod.cpp.o [341/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/schema_type_parser.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/schema_type_parser.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [342/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/torch/csrc/jit/frontend/source_range.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/serialization/source_range_serialization.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/csrc/jit/frontend/source_range.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [343/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Activation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Activation.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Activation.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [344/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveAveragePooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_adaptive_avg_pool2d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveAveragePooling.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [345/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveMaxPooling2d.cpp.o [346/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveAveragePooling3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_adaptive_avg_pool3d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveAveragePooling3d.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveAveragePooling3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [347/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AdaptiveMaxPooling3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptiveMaxPooling3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [348/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AffineGridGenerator.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AffineGridGenerator.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [349/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AmpKernels.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_amp_foreach_non_finite_check_and_unscale.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AmpKernels.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [350/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AutogradComposite.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/alias.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AutogradComposite.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [351/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AveragePool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AveragePool2d.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [352/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/AveragePool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AveragePool3d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [353/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BatchLinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BatchLinearAlgebra.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [354/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BatchLinearAlgebraKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BatchLinearAlgebraKernel.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [355/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BinaryOps.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [356/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/BlasKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BlasKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [357/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Blas.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Blas.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Blas.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [358/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/CPUBlas.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Matmul.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUBlas.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/OpMathType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUBlas.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUBlas.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [359/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Bucketization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/BucketizationUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Bucketization.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Bucketization.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [360/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ChanelShuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ChanelShuffle.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [361/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/CPUFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/CPUFallback.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [362/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ComparisonUtils.cpp.o [363/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Col2Im.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Col2Im.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Col2Im.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [364/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Constraints.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Constraints.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [365/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ConvolutionMM2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_slow_conv2d_forward.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionMM2d.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionMM2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [366/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Convolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Convolution.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [367/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ConvolutionTBC.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionTBC.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [368/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ConvolutionMM3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ConvolutionMM3d.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [369/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Copy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Copy.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Copy.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [370/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Correlation.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Correlation.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [371/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Cross.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Cross.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [372/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/DilatedMaxPool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/DilatedMaxPool2d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [373/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/DispatchStub.cpp.o [374/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/DilatedMaxPool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/DilatedMaxPool3d.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [375/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Distance.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Distance.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [376/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Dropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Dropout.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [377/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Distributions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Distributions.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Distributions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [378/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Embedding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Embedding.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [379/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Fill.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Fill.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [380/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/EmbeddingBag.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/EmbeddingBag.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/EmbeddingBag.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [381/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FractionalMaxPool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FractionalMaxPool2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [382/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ForeachOpsKernels.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/copy.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ForeachOpsKernels.cpp:57: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ForeachOpsKernels.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [383/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FunctionOfAMatrixUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FunctionOfAMatrixUtils.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [384/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FractionalMaxPool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FractionalMaxPool3d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FractionalMaxPool3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [385/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FusedAdagrad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fused_adagrad.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FusedAdagrad.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [386/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FusedAdam.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fused_adam.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FusedAdam.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [387/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/FusedSGD.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fused_sgd.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/FusedSGD.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [388/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/GatedLinearUnit.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/GatedLinearUnit.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [389/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/GridSampler.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/GridSampler.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [390/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Histogram.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Histogram.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [391/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/IndexingUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/IndexingUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/IndexingUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [392/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Im2Col.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Im2Col.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [393/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Integration.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Integration.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [394/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Itertools.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Itertools.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [395/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LegacyBridge.cpp.o [396/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Lerp.cpp.o [397/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LegacyBatching.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LegacyBatching.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [398/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Linear.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [399/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Loss.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Loss.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Loss.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [400/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebra.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebra.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [401/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossMultiLabelMargin.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossMultiLabelMargin.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [402/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossCTC.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossCTC.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [403/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossMultiMargin.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossMultiMargin.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [404/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossNLL.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [405/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/LossNLL2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL2d.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LossNLL2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [406/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/MaxPooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MaxPooling.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [407/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/MaxUnpooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MaxUnpooling.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [408/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Memory.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_pin_memory.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Memory.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [409/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/MetaTensor.cpp.o [410/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NNPACK.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NNPACK.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [411/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NaiveConvolutionTranspose2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose2d.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [412/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NaiveConvolutionTranspose3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose3d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveConvolutionTranspose3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [413/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NamedTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NamedTensor.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [414/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NaiveDilatedConvolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveDilatedConvolution.cpp:18: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NaiveDilatedConvolution.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [415/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/NegateFallback.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/MathBitsFallback.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NegateFallback.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [416/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Onehot.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Onehot.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [417/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PackedSequence.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/PackedSequence.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [418/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Normalization.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Normalization.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [419/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PadNd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/PadNd.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [420/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PointwiseOps.cpp.o [421/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/PixelShuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/roll.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorTransformations.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/PixelShuffle.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [422/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Pooling.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [423/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Pow.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Pow.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [424/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/QuantizedLinear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtilsMulti.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/QuantizedLinear.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [425/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/RangeFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/linspace.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [426/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReduceAllOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceAllOps.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [427/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/RNN.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RNN.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RNN.cpp:15: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [428/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReflectionPad.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReflectionPad.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [429/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReduceOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOps.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [430/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Repeat.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Repeat.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Repeat.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [431/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ReplicationPadding.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReplicationPadding.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [432/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/RowwisePrune.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RowwisePrune.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [433/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Resize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [434/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Scalar.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_local_scalar_dense.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Scalar.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Scalar.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [435/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SobolEngineOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOpsUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [436/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SegmentReduce.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SegmentReduce.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReductionType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SegmentReduce.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SegmentReduce.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [437/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SobolEngineOpsUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOpsUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SobolEngineOpsUtils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [438/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SoftMax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SoftMax.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SoftMax.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [439/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Sorting.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Sorting.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Sorting.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [440/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SparseTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [441/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SpectralOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SpectralOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [442/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/SummaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/zeros.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SummaryOps.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [443/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorCompare.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorCompare.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorCompare.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [444/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorAdvancedIndexing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.cpp:51: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [445/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorConversions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorConversions.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [446/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorIteratorReduce.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorIteratorReduce.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [447/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [448/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorProperties.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorProperties.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [449/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorTransformations.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/roll.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorTransformations.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorTransformations.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [450/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TestOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ScalarOps.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TestOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [451/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TriangularOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TriangularOpsUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TriangularOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [452/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TensorShape.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorShape.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [453/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/TypeProperties.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TypeProperties.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [454/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Unfold2d.cpp.o [455/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnaryOps.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [456/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Unfold3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Unfold3d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [457/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UnfoldBackward.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [458/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSample.cpp.o [459/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleBicubic2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_bicubic2d_aa.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleBicubic2d.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleBicubic2d.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [460/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/Unique.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Unique.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Unique.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [461/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleBilinear2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_bilinear2d_aa.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleBilinear2d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [462/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleLinear1d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/upsample_linear1d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleLinear1d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [463/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleNearest1d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_nearest_exact1d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleNearest1d.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [464/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleNearest2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_nearest_exact2d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleNearest2d.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [465/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleNearest3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_upsample_nearest_exact3d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleNearest3d.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [466/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/UpSampleTrilinear3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/upsample_trilinear3d.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UpSampleTrilinear3d.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [467/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/VariableMethodStubs.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/alias.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/VariableMethodStubs.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [468/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/WeightNorm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/WeightNorm.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [469/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/group_norm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/group_norm.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [470/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/layer_norm.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [471/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/verbose_wrapper.cpp.o [472/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/prim_native_functions.cpp.o [473/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/library.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/library.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [474/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/fbgemm_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/fbgemm_utils.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [475/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [476/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_dynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_dynamic.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [477/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_deserialize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_deserialize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [478/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_prepack.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [479/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_serialize.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_serialize.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [480/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/builtin_function.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ao_sparse/quantized/cpu/qlinear_unpack.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [481/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/ParamUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/ParamUtils.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [482/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/FlattenIndicesKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/FlattenIndicesCommon.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/FlattenIndicesKernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [483/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SoftMax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SoftMax.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [484/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseBlas.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBlas.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [485/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionCommon.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionKernel.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBinaryOpIntersectionCommon.h:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [486/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseBlasImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBlasImpl.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseBlasImpl.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [487/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseCsrTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseCsrTensor.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [488/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_unique.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseFactories.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [489/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseMatMul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseMatMul.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [490/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseCsrTensorMath.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseCsrTensorMath.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseCsrTensorMath.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [491/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseTensor.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseTensor.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [492/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseUnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseUnaryOps.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [493/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/ValidateCompressedIndicesKernel.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/SparseTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/ValidateCompressedIndicesCommon.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/ValidateCompressedIndicesKernel.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [494/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/sparse/SparseTensorMath.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorIndexing.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/sparse/SparseTensorMath.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [495/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorAliases.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorAliases.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [496/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorBackward.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorBackward.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [497/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorBinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorBinaryOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [498/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorFactories.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [499/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorMatmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUtils.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMatmul.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [500/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorMath.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMath.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NestedTensorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMath.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorMath.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [501/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorUnaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUnaryOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [502/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorTransformerFunctions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorTransformerFunctions.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [503/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/AffineQuantizer.cpp.o [504/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/AffineQuantizerBase.cpp.o [505/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/nested/NestedTensorUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/cat.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUtils.h:17, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/nested/NestedTensorUtils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [506/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/Copy.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/Copy.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [507/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/FakeQuantPerChannelAffine.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/FakeQuantPerChannelAffine.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [508/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/FakeQuantPerTensorAffine.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/FakeQuantPerTensorAffine.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [509/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/QTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/QTensor.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [510/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/TensorAdvancedIndexing.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/TensorAdvancedIndexing.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [511/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/TensorCompare.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/TensorCompare.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [512/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/AdaptiveAveragePooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/AdaptiveAveragePooling.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [513/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/TensorFactories.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/TensorFactories.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [514/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/AveragePool2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/AveragePool2d.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [515/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/AveragePool3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/AveragePool3d.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [516/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/ChannelShuffle.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/ChannelShuffle.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [517/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/BinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/BinaryOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/BinaryOps.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [518/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/IntReprQuant.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/IntReprQuant.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [519/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/LinearUnpackImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpp_custom_type_hack.h:57, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/LinearUnpackImpl.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/LinearUnpackImpl.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [520/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/MakePerTensorQuantizedTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/MakePerTensorQuantizedTensor.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [521/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/Normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/Normalization.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [522/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/ReduceOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/ReduceOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [523/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/RuyUtils.cpp.o [524/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/Pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/Pooling.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [525/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/Sorting.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/Sorting.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [526/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/TensorOperators.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/TensorOperators.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [527/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/UpSampleBilinear2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/UpSampleBilinear2d.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [528/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/TensorShape.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/TensorShape.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/TensorShape.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [529/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/UpSampleNearest2d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/UpSampleNearest2d.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [530/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/UpSampleNearest3d.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_empty_affine_quantized.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/UpSampleNearest3d.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [531/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/XnnpackUtils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/XnnpackUtils.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [532/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:21: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h: In instantiation of ‘c10::intrusive_ptr > deserialize_conv(ConvParamsSerializationTypeV3) [with unsigned int kSpatialDim = 2; ConvParamsSerializationTypeV3 = std::tuple >, std::vector, std::allocator > > >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:410:47: required from ‘int register_conv_params() [with int kSpatialDim = 2]’ 410 | return deserialize_conv(state); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:429:39: required from here 429 | TORCH_API int register_conv_params<2>(); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:316:11: warning: unused variable ‘groups’ [-Wunused-variable] 316 | int64_t groups = config_vals.at(idx); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:326:8: warning: unused variable ‘transpose’ [-Wunused-variable] 326 | bool transpose = flags & (1 << 0); | ^~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h: In instantiation of ‘c10::intrusive_ptr > deserialize_conv(ConvParamsSerializationTypeV3) [with unsigned int kSpatialDim = 3; ConvParamsSerializationTypeV3 = std::tuple >, std::vector, std::allocator > > >]’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:410:47: required from ‘int register_conv_params() [with int kSpatialDim = 3]’ 410 | return deserialize_conv(state); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.cpp:431:39: required from here 431 | TORCH_API int register_conv_params<3>(); | ^ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:316:11: warning: unused variable ‘groups’ [-Wunused-variable] 316 | int64_t groups = config_vals.at(idx); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/conv_serialization.h:326:8: warning: unused variable ‘transpose’ [-Wunused-variable] 326 | bool transpose = flags & (1 << 0); | ^~~~~~~~~ [533/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/init_qnnpack.cpp.o [534/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/fused_obs_fake_quant.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/_fake_quantize_per_tensor_affine_cachemask_tensor_qparams.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fused_obs_fake_quant.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [535/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qclamp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qclamp.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [536/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv.cpp:20: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [537/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv_dynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_dynamic.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_dynamic.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [538/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_prepack.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_prepack.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [539/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qconv_unpack_impl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_unpack_impl.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qconv_unpack_impl.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [540/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qdropout.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qdropout.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qdropout.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [541/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qelu.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qelu.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [542/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qembeddingbag.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/EmbeddingPackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [543/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qembeddingbag_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/EmbeddingPackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_prepack.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_prepack.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [544/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qgelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qgelu.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [545/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qembeddingbag_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/EmbeddingPackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_unpack.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qembeddingbag_unpack.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [546/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qhardsigmoid.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qhardsigmoid.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [547/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qhardswish.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qhardswish.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [548/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qlinear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [549/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qlinear_dynamic.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/PackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_dynamic.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_dynamic.cpp:11: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [550/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qlinear_prepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/PackedParams.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/fbgemm_utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_prepack.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qlinear_prepack.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [551/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qmatmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmatmul.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmatmul.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [552/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmul.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qmul.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [553/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qnormalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qnormalization.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qnormalization.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [554/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qrelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qrelu.cpp:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qrelu.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [555/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qsigmoid.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qsigmoid.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [556/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qtanh.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/QuantizedOps.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qtanh.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [557/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qsoftmax.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qsoftmax.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qsoftmax.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [558/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/qthreshold.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/qthreshold.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [559/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/library.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/library.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [560/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/qconv_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qconv_unpack.cpp:14: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qconv_unpack.cpp:16: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [561/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/LinearAlgebra.cpp.o [562/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/SparseBlasImpl.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SparseBlasImpl.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [563/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/SparseCsrLinearAlgebra.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseTensorImpl.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/SparseCsrTensorUtils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SparseCsrLinearAlgebra.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SparseCsrLinearAlgebra.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [564/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/qlinear_unpack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qlinear_unpack.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/qlinear_unpack.cpp:13: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [565/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/BinaryOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/BinaryOps.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [566/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Conv.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Conv.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [567/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/ConvPrepack.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Common.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/ConvPrepack.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [568/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Copy.cpp.o [569/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkl/SpectralOps.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkl/SpectralOps.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [570/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/IDeepRegistration.cpp.o [571/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Gelu.cpp.o [572/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Linear.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [573/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/MKLDNNCommon.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/MKLDNNCommon.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [574/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/MKLDNNConversions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/MKLDNNConversions.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/MKLDNNConversions.cpp:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [575/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Matmul.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/LinearAlgebraUtils.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Matmul.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Matmul.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [576/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/MkldnnTensorMath.cpp.o [577/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Normalization.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Normalization.cpp:17: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [578/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Pooling.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Pooling.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [579/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/OpContext.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/OpContext.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/OpContext.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [580/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Prelu.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Prelu.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [581/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/RegisterMkldnnOpContextClass.cpp.o [582/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Relu.cpp.o [583/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/SoftMax.cpp.o [584/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/TensorFactories.cpp.o [585/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/RNN.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/RNN.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/RNN.cpp:12: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [586/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/TensorShape.cpp.o [587/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/UnaryOps.cpp.o [588/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/mkldnn/Utils.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/mkldnn/Utils.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [589/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/transformers/attention.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty_like.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorOperators.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/attention.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [590/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/transformers/sdp_utils_cpp.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorSubclassLikeUtils.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/sdp_utils_cpp.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [591/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/utils/Factory.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/TensorNames.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/NamedTensorUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/utils/Factory.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [592/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/transformers/transformer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/transformers/transformer.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [593/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Activation.cpp.o [594/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/AveragePooling.cpp.o [595/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/ChannelShuffle.cpp.o [596/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Init.cpp.o [597/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Convolution.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Convolution.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Convolution.cpp:9: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [598/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Linear.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Linear.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Linear.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [599/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/MaxPooling.cpp.o [600/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/OpContext.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/Convolution.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/OpContext.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [601/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/Shim.cpp.o [602/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/xnnpack/RegisterOpContextClass.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/xnnpack/RegisterOpContextClass.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [603/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/CompositeViewCopyKernels.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/CompositeViewCopyKernels.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [604/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Functions.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.h:60, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Functions.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [605/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_0.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [606/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_1.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [607/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_3.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_3.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [608/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_2.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [609/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterBackendSelect.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterBackendSelect.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [610/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/Operators_4.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/stack.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/boxing/KernelFunction.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/Operators_4.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [611/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [612/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeExplicitAutograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutograd.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutograd.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [613/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeExplicitAutogradNonFunctional.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutogradNonFunctional.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeExplicitAutogradNonFunctional.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [614/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeImplicitAutogradNestedTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutogradNestedTensor.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutogradNestedTensor.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [615/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterCompositeImplicitAutograd.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutograd.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterCompositeImplicitAutograd.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [616/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_0.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_0.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [617/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_1.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_1.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [618/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_2.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_2.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [619/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterFunctionalization_3.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterFunctionalization_3.cpp:6: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/FunctionalTensorWrapper.h:10: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [620/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterMkldnnCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMkldnnCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMkldnnCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [621/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterNestedTensorCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [622/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterNestedTensorMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterNestedTensorMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [623/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterQuantizedCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [624/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [625/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterQuantizedMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterQuantizedMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [626/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSchema.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/function_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/op_registration/infer_schema.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:61, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSchema.cpp:3: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [627/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [628/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseCsrCPU.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrCPU.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrCPU.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [629/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseCsrMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseCsrMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [630/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterSparseMeta.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseMeta.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterSparseMeta.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [631/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/UfuncCPU_add.cpp.o [632/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/RegisterZeroTensor.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterZeroTensor.cpp:28: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/RegisterZeroTensor.cpp:47: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [633/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/ATenOpList.cpp.o [634/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/quantized/QTensorImpl.cpp.o [635/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/core/TensorMethods.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorMethods.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [636/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_bind.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.cpp:5: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [637/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/quantized/Quantizer.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/quantized/Quantizer.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [638/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_model_loader.cpp.o [639/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_wrapper.cpp.o [640/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/UfuncCPUKernel_add.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/AccumulateType.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Math.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec_base.h:27, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec256/vec256.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/vec.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional_base.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/cpu/vec/functional.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ufunc/add.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/UfuncCPUKernel_add.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/UfuncCPUKernel_add.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [641/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/nnapi/nnapi_register.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ATen.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_register.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/OperatorEntry.h:12, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/Dispatcher.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/library.h:63, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/torch/custom_class.h:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/nnapi/nnapi_bind.h:7: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h: In lambda function: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:32: warning: possibly dangling reference to a temporary [-Wdangling-reference] 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/dispatch/DispatchKeyExtractor.h:154:61: note: the temporary was destroyed at the end of the full expression ‘__for_begin .c10::impl::ListIterator > >::operator*().c10::impl::ListElementReference > >::operator std::conditional_t()’ 154 | for (const at::Tensor& tensor : ivalue.toTensorList()) { | ^ [642/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/spherical_bessel_j0.cpp.DEFAULT.cpp.o [643/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k1.cpp.DEFAULT.cpp.o [644/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/scaled_modified_bessel_k0.cpp.DEFAULT.cpp.o [645/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:19, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/layer_norm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/layer_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/layer_norm_kernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [646/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int8mm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int8mm_kernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int8mm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [647/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/quantized/cpu/kernels/QuantizedOpKernels.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [648/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/int4mm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/int4mm_kernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/int4mm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [649/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:20, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/batch_norm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/batch_norm.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/batch_norm_kernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [650/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/airy_ai.cpp.DEFAULT.cpp.o [651/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/group_norm_kernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:21, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/group_norm_kernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/group_norm_kernel.cpp:8: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [652/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/WeightNormKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/WeightNormKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/WeightNormKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [653/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleMoreKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [654/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/arange.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnfoldBackward.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnfoldBackwardKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [655/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UpSampleKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UpSampleKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UpSampleKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [656/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/Unfold2d.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/Unfold2d.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/Unfold2d.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [657/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/WrapDimUtils.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/TensorCompareKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/TensorCompareKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [658/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SumKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/AccumulateType.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SumKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SumKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [659/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/StackKernel.cpp.DEFAULT.cpp.o [660/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SpmmReduceKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [661/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Generator.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/UnaryOps.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/UnaryOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [662/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SparseFactories.cpp.DEFAULT.cpp.o [663/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SoftMaxKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SoftMaxKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [664/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SortingKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SortingKernel.cpp:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SortingKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [665/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue_inl.h:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/ivalue.h:1581, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List_inl.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/List.h:490, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorAdvancedIndexing.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/TensorBase.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/NonEmptyUtils.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ScatterGatherKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [666/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RenormKernel.cpp.DEFAULT.cpp.o [667/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/view.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/ExpandUtils.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/SampledAddmmKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [668/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ReduceOpsUtils.h:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:5, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceAllOpsKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [669/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/RangeFactories.h:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/RangeFactoriesKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [670/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/scalar_tensor.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/TensorFactories.h:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/ResizeCommon.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Resize.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/ReduceOpsKernel.cpp:4: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [671/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PowKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PowKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PowKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [672/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PixelShuffleKernel.cpp.DEFAULT.cpp.o [673/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PointwiseOpsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [674/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/NativeMultiheadAttnKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [675/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MultinomialKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:15, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MultinomialKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MultinomialKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [676/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxUnpoolKernel.cpp.DEFAULT.cpp.o [677/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPooling.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MaxPooling.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MaxPooling.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [678/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/PaddingKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/PaddingKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/PaddingKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [679/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/AdaptivePooling.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/MaxPoolKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/MaxPoolKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [680/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/LinearAlgebraKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [681/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/LerpKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/OpMathType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Lerp.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/LerpKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/LerpKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [682/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/HistogramKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/aminmax.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:13, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/HistogramKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/Histogram.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/HistogramKernel.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [683/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/IndexKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Context.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/IndexKernel.cpp:7, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/IndexKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [684/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedSGDKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedSGDKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [685/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedAdamKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedAdamKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [686/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FusedAdagradKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [687/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FunctionOfAMatrixUtilsKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [688/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/GridSamplerKernel.cpp.DEFAULT.cpp.o [689/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/FillKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/DeprecatedTypeProperties.h:4, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/Dispatch_v2.h:1, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/FillKernel.cpp:2, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/FillKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [690/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistributionKernels.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:14, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DistributionKernels.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/TensorImpl.h:11, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/GeneratorImpl.h:8, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Generator.h:18, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/CPUGeneratorImpl.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DistributionKernels.cpp:2: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~ [691/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DistanceOpsKernel.cpp.DEFAULT.cpp.o [692/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef.h:631, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/DeviceGuard.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/ops/empty.h:6, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp:10, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/DepthwiseConvKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h: In static member function ‘static c10::detail::IListRefConstRef c10::detail::IListRefTagImpl::iterator_get(const c10::List >::const_iterator&)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:17: warning: possibly dangling reference to a temporary [-Wdangling-reference] 171 | const auto& ivalue = (*it).get(); | ^~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/IListRef_inl.h:171:35: note: the temporary was destroyed at the end of the full expression ‘(& it)->c10::impl::ListIterator, __gnu_cxx::__normal_iterator > >::operator*().c10::impl::ListElementReference, __gnu_cxx::__normal_iterator > >::get()’ 171 | const auto& ivalue = (*it).get(); | ~~~~~~~~~^~ [693/1912] Building CXX object caffe2/CMakeFiles/torch_cpu.dir/__/aten/src/ATen/native/cpu/CrossKernel.cpp.DEFAULT.cpp.o In file included from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/ScalarType.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/core/Scalar.h:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/core/TensorBody.h:16, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/core/Tensor.h:3, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/aten/src/ATen/native/cpu/CrossKernel.cpp:9, from /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/build/aten/src/ATen/native/cpu/CrossKernel.cpp.DEFAULT.cpp:1: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h: In function ‘float c10::detail::f32_from_bits(uint16_t)’: /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:40:7: warning: ‘tmp’ is used uninitialized [-Wuninitialized] 40 | res = *tempRes; | ~~~~^~~~~~~~~~ /builddir/build/BUILD/python-torch-2.4.0-build/pytorch-v2.4.0/c10/util/BFloat16.h:31:12: note: ‘tmp’ declared here 31 | uint32_t tmp = src; | ^~~